Merge branch 'main' of hf.co:spaces/Awlly/NLP_app
Browse files- README.md +79 -0
- __pycache__/preprocessing.cpython-310.pyc +0 -0
- app_models/__pycache__/bag_of_words_MODEL.cpython-310.pyc +0 -0
- app_models/__pycache__/gpt_MODEL.cpython-310.pyc +0 -0
- app_models/__pycache__/lstm_MODEL.cpython-310.pyc +0 -0
- app_models/__pycache__/rubert_MODEL.cpython-310.pyc +0 -0
- app_models/__pycache__/toxicity_MODEL.cpython-310.pyc +0 -0
- app_models/gpt_MODEL.py +2 -2
- app_pages/__pycache__/page1_model_comparison.cpython-310.pyc +0 -0
- app_pages/__pycache__/page2_rubert_toxicity.cpython-310.pyc +0 -0
- app_pages/__pycache__/page3_gpt_model.cpython-310.pyc +0 -0
- app_pages/page1_model_comparison.py +29 -7
- app_pages/page3_gpt_model.py +3 -2
README.md
CHANGED
|
@@ -10,3 +10,82 @@ pinned: false
|
|
| 10 |
---
|
| 11 |
|
| 12 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
---
|
| 11 |
|
| 12 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
| 13 |
+
|
| 14 |
+
## NLP App Hugging Face's logo
|
| 15 |
+
Hugging Face
|
| 16 |
+
# Streamlit app with computer vision 💡
|
| 17 |
+
Elbrus Bootcamp | Phase-2 | Team Project
|
| 18 |
+
|
| 19 |
+
## Team🧑🏻💻
|
| 20 |
+
1. [Awlly](https://github.com/Awlly)
|
| 21 |
+
2. [sakoser](https://github.com/sakoser)
|
| 22 |
+
3. [whoisida]https://github.com/whoisida
|
| 23 |
+
|
| 24 |
+
## Task 📌lassifi
|
| 25 |
+
Create a service that classifies movie reviews into good, neutral and bad categories, a service that classifies user input as toxic or non-toxic, as well as a GPT 2 based text generation service that was trained to emulate a certain author’s writing.
|
| 26 |
+
|
| 27 |
+
## Contents 📝
|
| 28 |
+
1. Classifies movie reviewsusing LSTM,ruBert,BOW 💨 [Dataset](https://drive.google.com/file/d/1c92sz81bEfOw-rutglKpmKGm6rySmYbt/view?usp=sharing)
|
| 29 |
+
2. classifies user input as toxic or non-toxi using ruBert-tiny-toxicity 📑 [Dataset](https://drive.google.com/file/d/1O7orH9CrNEhnbnA5KjXji8sgrn6iD5n-/view?usp=drive_link)
|
| 30 |
+
3. GPT 2 based text generation service
|
| 31 |
+
|
| 32 |
+
## Deployment 🎈
|
| 33 |
+
The service is implemented on [Hugging Face](https://huggingface.co/spaces/Awlly/NLP_app)
|
| 34 |
+
|
| 35 |
+
## Libraries 📖
|
| 36 |
+
```python
|
| 37 |
+
import os
|
| 38 |
+
import unicodedata
|
| 39 |
+
import nltk
|
| 40 |
+
from dataclasses import dataclass
|
| 41 |
+
import joblib
|
| 42 |
+
import numpy as np
|
| 43 |
+
import matplotlib.pyplot as plt
|
| 44 |
+
import torch
|
| 45 |
+
import torch.nn as nn
|
| 46 |
+
import torch.nn.functional as F
|
| 47 |
+
import torch.optim as optim
|
| 48 |
+
from torch.utils.data import DataLoader, TensorDataset
|
| 49 |
+
from torchvision.datasets import ImageFolder
|
| 50 |
+
from torchvision import datasets
|
| 51 |
+
from torchvision import transforms as T
|
| 52 |
+
from torchvision.io import read_image
|
| 53 |
+
from torch.utils.data import Dataset, random_split
|
| 54 |
+
import torchutils as tu
|
| 55 |
+
from transformers import GPT2LMHeadModel, GPT2Tokenizer
|
| 56 |
+
from typing import Tuple
|
| 57 |
+
from tqdm import tqdm
|
| 58 |
+
from transformers import AutoModel, AutoTokenizer
|
| 59 |
+
from transformers import AutoModelForSequenceClassification
|
| 60 |
+
import pydensecrf.densecrf as dcrf
|
| 61 |
+
import pydensecrf.utils as dcrf_utils
|
| 62 |
+
from preprocessing import data_preprocessing
|
| 63 |
+
import streamlit as st
|
| 64 |
+
import string
|
| 65 |
+
from sklearn.linear_model import LogisticRegression
|
| 66 |
+
import re
|
| 67 |
+
|
| 68 |
+
|
| 69 |
+
|
| 70 |
+
|
| 71 |
+
from preprocessing import preprocess_single_string
|
| 72 |
+
```
|
| 73 |
+
|
| 74 |
+
|
| 75 |
+
from preprocessing import data_preprocessing
|
| 76 |
+
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
|
| 80 |
+
## Guide 📜
|
| 81 |
+
#### How to run locally?
|
| 82 |
+
|
| 83 |
+
1. To create a Python virtual environment for running the code, enter:
|
| 84 |
+
|
| 85 |
+
``python3 -m venv my-env``
|
| 86 |
+
|
| 87 |
+
2. Activate the new environment:
|
| 88 |
+
|
| 89 |
+
* Windows: ```my-env\Scripts\activate.bat```
|
| 90 |
+
* macOS and Linux: ```source my-env/bin/activate```
|
| 91 |
+
|
__pycache__/preprocessing.cpython-310.pyc
DELETED
|
Binary file (2.32 kB)
|
|
|
app_models/__pycache__/bag_of_words_MODEL.cpython-310.pyc
DELETED
|
Binary file (630 Bytes)
|
|
|
app_models/__pycache__/gpt_MODEL.cpython-310.pyc
DELETED
|
Binary file (1.08 kB)
|
|
|
app_models/__pycache__/lstm_MODEL.cpython-310.pyc
DELETED
|
Binary file (3.49 kB)
|
|
|
app_models/__pycache__/rubert_MODEL.cpython-310.pyc
DELETED
|
Binary file (1.43 kB)
|
|
|
app_models/__pycache__/toxicity_MODEL.cpython-310.pyc
DELETED
|
Binary file (985 Bytes)
|
|
|
app_models/gpt_MODEL.py
CHANGED
|
@@ -10,7 +10,7 @@ model = GPT2LMHeadModel.from_pretrained(model_path)
|
|
| 10 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 11 |
model.to(device)
|
| 12 |
|
| 13 |
-
def generate_text(prompt_text, length, temperature):
|
| 14 |
encoded_prompt = tokenizer.encode(prompt_text, add_special_tokens=False, return_tensors="pt")
|
| 15 |
encoded_prompt = encoded_prompt.to(device)
|
| 16 |
|
|
@@ -22,7 +22,7 @@ def generate_text(prompt_text, length, temperature):
|
|
| 22 |
top_p=0.9,
|
| 23 |
repetition_penalty=1.2,
|
| 24 |
do_sample=True,
|
| 25 |
-
num_return_sequences=
|
| 26 |
)
|
| 27 |
|
| 28 |
# Decode the generated text
|
|
|
|
| 10 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 11 |
model.to(device)
|
| 12 |
|
| 13 |
+
def generate_text(prompt_text, length, temperature, beams):
|
| 14 |
encoded_prompt = tokenizer.encode(prompt_text, add_special_tokens=False, return_tensors="pt")
|
| 15 |
encoded_prompt = encoded_prompt.to(device)
|
| 16 |
|
|
|
|
| 22 |
top_p=0.9,
|
| 23 |
repetition_penalty=1.2,
|
| 24 |
do_sample=True,
|
| 25 |
+
num_return_sequences=beams,
|
| 26 |
)
|
| 27 |
|
| 28 |
# Decode the generated text
|
app_pages/__pycache__/page1_model_comparison.cpython-310.pyc
DELETED
|
Binary file (904 Bytes)
|
|
|
app_pages/__pycache__/page2_rubert_toxicity.cpython-310.pyc
DELETED
|
Binary file (794 Bytes)
|
|
|
app_pages/__pycache__/page3_gpt_model.cpython-310.pyc
DELETED
|
Binary file (845 Bytes)
|
|
|
app_pages/page1_model_comparison.py
CHANGED
|
@@ -2,6 +2,7 @@ import streamlit as st
|
|
| 2 |
from app_models.rubert_MODEL import classify_text
|
| 3 |
from app_models.bag_of_words_MODEL import predict
|
| 4 |
from app_models.lstm_MODEL import predict_review
|
|
|
|
| 5 |
|
| 6 |
class_prefix = 'This review is likely...'
|
| 7 |
|
|
@@ -11,11 +12,32 @@ def run():
|
|
| 11 |
|
| 12 |
# Example placeholder for user input
|
| 13 |
user_input = st.text_area("")
|
| 14 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
# Placeholder buttons for model selection
|
| 16 |
-
if st.button('Classify with BoW/TF-IDF'):
|
| 17 |
-
|
| 18 |
-
if st.button('Classify with LSTM'):
|
| 19 |
-
|
| 20 |
-
if st.button('Classify with ruBERT'):
|
| 21 |
-
|
|
|
|
| 2 |
from app_models.rubert_MODEL import classify_text
|
| 3 |
from app_models.bag_of_words_MODEL import predict
|
| 4 |
from app_models.lstm_MODEL import predict_review
|
| 5 |
+
import time
|
| 6 |
|
| 7 |
class_prefix = 'This review is likely...'
|
| 8 |
|
|
|
|
| 12 |
|
| 13 |
# Example placeholder for user input
|
| 14 |
user_input = st.text_area("")
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
if st.button('Classify with All Models'):
|
| 18 |
+
# Measure and display Bag of Words/TF-IDF prediction time
|
| 19 |
+
start_time = time.time()
|
| 20 |
+
bow_tfidf_result = predict(user_input)
|
| 21 |
+
end_time = time.time()
|
| 22 |
+
st.write(f'{class_prefix} {bow_tfidf_result} according to Bag of Words/TF-IDF. Time taken: {end_time - start_time:.2f} seconds.')
|
| 23 |
+
|
| 24 |
+
# Measure and display LSTM prediction time
|
| 25 |
+
start_time = time.time()
|
| 26 |
+
lstm_result = predict_review(user_input)
|
| 27 |
+
end_time = time.time()
|
| 28 |
+
st.write(f'{class_prefix} {lstm_result} according to LSTM. Time taken: {end_time - start_time:.2f} seconds.')
|
| 29 |
+
|
| 30 |
+
# Measure and display ruBERT prediction time
|
| 31 |
+
start_time = time.time()
|
| 32 |
+
rubert_result = classify_text(user_input)
|
| 33 |
+
end_time = time.time()
|
| 34 |
+
st.write(f'{class_prefix} {rubert_result} according to ruBERT. Time taken: {end_time - start_time:.2f} seconds.')
|
| 35 |
+
|
| 36 |
+
|
| 37 |
# Placeholder buttons for model selection
|
| 38 |
+
# if st.button('Classify with BoW/TF-IDF'):
|
| 39 |
+
# st.write(f'{class_prefix}{predict(user_input)}')
|
| 40 |
+
# if st.button('Classify with LSTM'):
|
| 41 |
+
# st.write(f'{class_prefix}{predict_review(user_input)}')
|
| 42 |
+
# if st.button('Classify with ruBERT'):
|
| 43 |
+
# st.write(f'{class_prefix}{classify_text(user_input)}')
|
app_pages/page3_gpt_model.py
CHANGED
|
@@ -6,9 +6,10 @@ def run():
|
|
| 6 |
st.title('GPT Text Generation')
|
| 7 |
prompt_text = st.text_area("Input Text", "Type here...")
|
| 8 |
length = st.slider("Length of Generated Text", min_value=50, max_value=500, value=200)
|
| 9 |
-
temperature = st.slider("Temperature", min_value=0.1, max_value=
|
|
|
|
| 10 |
|
| 11 |
if st.button('Generate Text'):
|
| 12 |
with st.spinner('Generating...'):
|
| 13 |
-
generated_text = generate_text(prompt_text, length, temperature)
|
| 14 |
st.text_area("Generated Text", generated_text, height=250)
|
|
|
|
| 6 |
st.title('GPT Text Generation')
|
| 7 |
prompt_text = st.text_area("Input Text", "Type here...")
|
| 8 |
length = st.slider("Length of Generated Text", min_value=50, max_value=500, value=200)
|
| 9 |
+
temperature = st.slider("Temperature", min_value=0.1, max_value=2.0, value=0.7, step=0.1)
|
| 10 |
+
beams = st.slider("Number of Generations", min_value=2, max_value=10, value=4, step=1)
|
| 11 |
|
| 12 |
if st.button('Generate Text'):
|
| 13 |
with st.spinner('Generating...'):
|
| 14 |
+
generated_text = generate_text(prompt_text, length, temperature, beams)
|
| 15 |
st.text_area("Generated Text", generated_text, height=250)
|