Update README.md
Browse files
README.md
CHANGED
|
@@ -61,8 +61,8 @@ This model is designed for extractive question answering tasks where:
|
|
| 61 |
from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
|
| 62 |
|
| 63 |
# Load model and tokenizer
|
| 64 |
-
model = AutoModelForQuestionAnswering.from_pretrained("
|
| 65 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
| 66 |
|
| 67 |
# Create QA pipeline
|
| 68 |
qa_pipeline = pipeline(
|
|
@@ -99,8 +99,8 @@ Confidence: 0.9234
|
|
| 99 |
import torch
|
| 100 |
from transformers import AutoModelForQuestionAnswering, AutoTokenizer
|
| 101 |
|
| 102 |
-
model = AutoModelForQuestionAnswering.from_pretrained("
|
| 103 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
| 104 |
|
| 105 |
question = "What is the capital of France?"
|
| 106 |
context = "Paris is the capital and largest city of France."
|
|
|
|
| 61 |
from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
|
| 62 |
|
| 63 |
# Load model and tokenizer
|
| 64 |
+
model = AutoModelForQuestionAnswering.from_pretrained("G20-CS4248/bert-baseline-qa")
|
| 65 |
+
tokenizer = AutoTokenizer.from_pretrained("G20-CS4248/bert-baseline-qa")
|
| 66 |
|
| 67 |
# Create QA pipeline
|
| 68 |
qa_pipeline = pipeline(
|
|
|
|
| 99 |
import torch
|
| 100 |
from transformers import AutoModelForQuestionAnswering, AutoTokenizer
|
| 101 |
|
| 102 |
+
model = AutoModelForQuestionAnswering.from_pretrained("G20-CS4248/bert-baseline-qa")
|
| 103 |
+
tokenizer = AutoTokenizer.from_pretrained("G20-CS4248/bert-baseline-qa")
|
| 104 |
|
| 105 |
question = "What is the capital of France?"
|
| 106 |
context = "Paris is the capital and largest city of France."
|