huggingface / autotrain-advanced Goto Github PK
View Code? Open in Web Editor NEW๐ค AutoTrain Advanced
Home Page: https://huggingface.co/autotrain
License: Apache License 2.0
๐ค AutoTrain Advanced
Home Page: https://huggingface.co/autotrain
License: Apache License 2.0
Hello,
--max_models is a required parameter. However, it doesn't appear in the task doc pages, e.g. https://huggingface.co/docs/autonlp/multi_class_classification.html
Thank you :)
We have finetuned our BERT model for text2text generation. It is working fine on the Jupyter notebook. But when I use the same trained model on another server of Ubuntu, then it shows the issue. This is my first post, so please bear with me. The issue I'm facing is that when I generate output on small sentences, it works fine. But on long sentences, it shows the following error:
At most 4 tokens in tensor([ 2, 2, 2, 2, 44763, 44763, 2, 44763]) can be equal to eos_token_id: 2. Make sure tensor([ 2, 2, 2, 2, 44763, 44763, 2, 44763]) are corrected.
My output generation code is:
from simpletransformers.seq2seq import Seq2SeqModel
#logging.basicConfig(level=logging.INFO)
#transformers_logger = logging.getLogger("transformers")
#transformers_logger.setLevel(logging.ERROR)
model = Seq2SeqModel(
encoder_decoder_type="bart", encoder_decoder_name="PATHOFMODEL",use_cuda=False,)
while True:
original = input("Enter text to paraphrase: ")
to_predict = [original]
preds = model.predict(to_predict)
for pred in preds[0]:
print(pred)
This code works fine on notebook server where I trained the model. But I install all the dependencies on simple ubuntu server and then run this code with the trained model files. It works for some sentences but not for some other sentences.
Here's the complete issue on StackOverflow.
https://stackoverflow.com/q/67195582
Someone told me to change the TensorFlow version. I tried it. It worked for one day. After that, I faced the same problem again.
minority sample are around ~300 in number
Any plans to implement this?
Hello,
Tried to train a model using AutoNLP and all 5 models failed. I verified the input CSVs are valid by loading them using datasets.
Any idea?
The project is called "appliances" and the owner is "mostrovsky"
Really anxious for this feature!
I tried to create a project with the given configuration:
project = client.create_project(name="project", task="single_column_regression", language="en", max_models=1, hub_model="EleutherAI/gpt-neo-2.7B")
and keep getting
---------------------------------------------------------------------------
HTTPError Traceback (most recent call last)
<ipython-input-23-f15ed7bdca2f> in <module>
----> 1 project = client.create_project(name="project", task="single_column_regression", language="en", max_models=1, hub_model="EleutherAI/gpt-neo-2.7B")
2
/usr/local/lib/python3.7/site-packages/autonlp/autonlp.py in create_project(self, name, task, language, max_models, hub_model)
95 },
96 }
---> 97 json_resp = http_post(path="/projects/create", payload=payload, token=self.token).json()
98 proj_name = json_resp["proj_name"]
99 created = json_resp["created"]
/usr/local/lib/python3.7/site-packages/autonlp/utils.py in http_post(path, token, payload, domain, suppress_logs, **kwargs)
64 except requests.exceptions.ConnectionError:
65 raise UnreachableAPIError("โ Failed to reach AutoNLP API, check your internet connection")
---> 66 response.raise_for_status()
67 return response
68
/usr/local/lib/python3.7/site-packages/requests/models.py in raise_for_status(self)
941
942 if http_error_msg:
--> 943 raise HTTPError(http_error_msg, response=self)
944
945 def close(self):
HTTPError: 400 Client Error: Bad Request for url: https://api.autonlp.huggingface.co/projects/create
also if I specify other models like gpt2
. It works when I set hub_model=""
, though. Is the problem the models I try to specify which do not support the single_column_regression
task?
When I am trying to upload training set from the CLI as per the instructions given in the README I get the following error:-
If not specifying `clone_from`, you need to pass Repository a valid git clone.
Traceback (most recent call last):
File "/Users/****/opt/anaconda3/bin/autonlp", line 8, in <module>
sys.exit(main())
File "/Users/****/opt/anaconda3/lib/python3.8/site-packages/autonlp/cli/autonlp.py", line 40, in main
command.run()
File "/Users/****/opt/anaconda3/lib/python3.8/site-packages/autonlp/cli/upload.py", line 109, in run
project.upload(filepaths=files, split=self._split, col_mapping=col_maps)
File "/Users/****/opt/anaconda3/lib/python3.8/site-packages/autonlp/project.py", line 171, in upload
dataset_repo = Repository(
File "/Users/****/opt/anaconda3/lib/python3.8/site-packages/huggingface_hub/repository.py", line 69, in __init__
raise ValueError(
ValueError: If not specifying `clone_from`, you need to pass Repository a valid git clone.
Getting the same error from the python API too.
Is it possible for the Arabic
language?
There is obviously the possibility of creating projects but is there a command to delete projects? Or where are those definitions stored (what is the location of the config file)
On the website it says that regression
is one of the available tasks. In the documentation it's not mentioned. Is it (already) available?
When attempting to upload a CSV training set for my model I receive a JSONDecodeError
error. I tried uploading my smaller validation set too, but it also failed. I'm not entirely sure why JSON decoders are even being ran against a CSV file.
At first I thought maybe the CSV was invalid, but it checks out. I am not sure how to debug this problem.
Any help is greatly appreciated! Thank you.
$ csvclean ~/training_set.csv
No errors.
col_one,col_two
TRUE,"Lorem ipsum dolor sit amet, consectetur adipiscing elit"
FALSE,"Ut id ex luctus ""with quoted text inside"" vitae tincidunt nibh"
TRUE,"Nam ligula nibh, dapibus eget justo vitae"
FALSE,"Cras sed molestie enim. Etiam facilisis erat id bibendum"
$ autonlp upload --project my_project \
--split train \
--col_mapping col_one:target,col_two:text \
--files ~/training_set.csv
> INFO Uploading files for project: my_project
> INFO ๐ Retrieving credentials from config...
> INFO โ Retrieving project 'my_project' from AutoNLP...
> INFO ๐ Refreshing project status...
> INFO ๐ Refreshing uploaded files information...
> INFO ๐ Refreshing models information...
> INFO ๐ Refreshing cost information...
> INFO โ
Successfully loaded project: 'my_project'!
> INFO Mapping: {'col_one': 'target', 'col_two': 'text'}
Traceback (most recent call last):
File "/usr/local/bin/autonlp", line 8, in <module>
sys.exit(main())
File "/usr/local/lib/python3.9/site-packages/autonlp/cli/autonlp.py", line 57, in main
details = err.response.json().get("detail")
File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 900, in json
return complexjson.loads(self.text, **kwargs)
File "/usr/local/Cellar/[email protected]/3.9.5/Frameworks/Python.framework/Versions/3.9/lib/python3.9/json/__init__.py", line 346, in loads
return _default_decoder.decode(s)
File "/usr/local/Cellar/[email protected]/3.9.5/Frameworks/Python.framework/Versions/3.9/lib/python3.9/json/decoder.py", line 337, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/usr/local/Cellar/[email protected]/3.9.5/Frameworks/Python.framework/Versions/3.9/lib/python3.9/json/decoder.py", line 355, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
$ autonlp --version
0.3.1
$ python -V
Python 3.9.5
$ pip -V
pip 21.1.3
INFO Fetching info for project: Summarization
INFO ๐ Retrieving credentials from config...
INFO โ Retrieving project 'Summarization' from AutoNLP...
INFO ๐ Refreshing project status...
INFO ๐ Refreshing uploaded files information...
INFO ๐ Refreshing models information...
INFO ๐ Refreshing cost information...
INFO โ Successfully loaded project: 'Summarization'!
AutoNLP Project (id # 2068)
โข Name: Summarization
โข Owner: hiiamsid
โข Status: โ Failed to upload processed data files to the huggingface hub
โข Task: Summarization
โข Created at: 2021-10-18 06:33 Z
โข Last update: 2021-10-18 10:55 Z
๐ฐ Project current cost: USD 481.79
~~~~~~~~~~~~~~ Files ~~~~~~~~~~~~~~
Dataset ID:
hiiamsid/autonlp-data-Summarization
๐ spanish_paraphrase_train.csv (id # 1775)
โข Split: train
โข Processing status: โ
Success!
โข Last update: 2021-10-18 10:44 Z
๐ es_paraphrase_test.csv (id # 1776)
โข Split: valid
โข Processing status: โ
Success!
โข Last update: 2021-10-18 10:43 Z
Users should be able to list their projects
Hi all, thanks for this amazing product!
I am currently testing various features on the trial version, but the Entity Extraction documentation page at https://huggingface.co/docs/autonlp/entity_extraction.html is currently empty. Is this feature not offered yet? Or can I find its documentation elsewhere? Thanks!
When uploading a dataset, upload.py
raises a TypeError
stating that get_project()
is missing is_eval
argument.
$ autonlp upload --project sentiment_detection --split train \
--col_mapping review:text,sentiment:target \
--files ~/datasets/train.csv
> INFO Uploading files for project: sentiment_detection
Traceback (most recent call last):
File "/Users/sb/.pyenv/versions/test/bin/autonlp", line 8, in <module>
sys.exit(main())
File "/Users/sb/.pyenv/versions/test/lib/python3.9/site-packages/autonlp/cli/autonlp.py", line 54, in main
command.run()
File "/Users/sb/.pyenv/versions/test/lib/python3.9/site-packages/autonlp/cli/upload.py", line 82, in run
project = client.get_project(name=self._name)
TypeError: get_project() missing 1 required positional argument: 'is_eval'
If I manually revise upload.py
as follows
project = client.get_project(name=self._name, is_eval=False)
things work as they should. Likewise, this happens when trying to run the train
command.
$ autonlp train --project sentiment_detection
> INFO Starting Training For Project: sentiment_detection
Traceback (most recent call last):
File "/Users/sb/.pyenv/versions/test/bin/autonlp", line 8, in <module>
sys.exit(main())
File "/Users/sb/.pyenv/versions/test/lib/python3.9/site-packages/autonlp/cli/autonlp.py", line 54, in main
command.run()
File "/Users/sb/.pyenv/versions/test/lib/python3.9/site-packages/autonlp/cli/train.py", line 34, in run
project = client.get_project(name=self._name)
TypeError: get_project() missing 1 required positional argument: 'is_eval'
I'm using autonlp
version 0.3.0.
$ autonlp --version
0.3.0
Is it possible to add the Malay Language into the service?
Will autonlp also train GPT-neo models for tasks like regression and classification?
Hindi language pipeline isn't available for coref
How to reproduce:
autonlp metrics --project <project_name>
Expected output:
2021-02-25 14:02:18.178 | INFO | autonlp.autonlp:_login_from_conf:67 - ๐ Retrieving credentials from config...
Traceback (most recent call last):
File "C:\Users\sbran\miniconda3\envs\autonlp-front\Scripts\autonlp-script.py", line 33, in <module>
sys.exit(load_entry_point('autonlp', 'console_scripts', 'autonlp')())
File "c:\users\sbran\documents\dev\huggingface\autonlp\src\autonlp\cli\autonlp.py", line 34, in main
command.run()
File "c:\users\sbran\documents\dev\huggingface\autonlp\src\autonlp\cli\metrics.py", line 33, in run
_ = client.get_metrics(model_id=self._model_id, project=self._project)
File "c:\users\sbran\documents\dev\huggingface\autonlp\src\autonlp\autonlp.py", line 146, in get_metrics
_metrics = Metrics.from_json_resp(
File "c:\users\sbran\documents\dev\huggingface\autonlp\src\autonlp\metrics.py", line 21, in from_json_resp
language=json_resp["config"]["language"],
KeyError: 'config'
This is due to fact we recently stripped config from project API response in the backend, in fact removing language information.
Parameter max_models not included in the create_project
function call in the Readme.
Suggested:
project = client.create_project(name="sentiment_detection", task="binary_classification", language="en", max_models=5)
Hi! This issue is most like a recommendation.
When I tried to upload a .csv file in huggingface, I don't have the option to choose a separator.
I understand that the .csv file is a "comma separated values" but in some cases the .csv has another separator, like in this case.
I think the problem it will resolve if I change the "|" to "," but hopefully, in the future, you can implement something like what I said above.
Best regards!
My project status is still created and not in active state. Unable to train the model.
When calling the "project.upload" function in python I get the following error:
I've substituted the AutoNLP api token with "API_TOKEN" and the folder/model name with MODEL_PATH. How can I fix it?
CalledProcessError Traceback (most recent call last)
~/opt/anaconda3/envs/aitoci_py37/lib/python3.7/site-packages/huggingface_hub/repository.py in clone_from(self, repo_url, use_auth_token)
147 encoding="utf-8",
--> 148 cwd=self.local_dir,
149 )
~/opt/anaconda3/envs/aitoci_py37/lib/python3.7/subprocess.py in run(input, capture_output, timeout, check, *popenargs, **kwargs)
511 raise CalledProcessError(retcode, process.args,
--> 512 output=stdout, stderr=stderr)
513 return CompletedProcess(process.args, retcode, stdout, stderr)
CalledProcessError: Command '['git', 'clone', 'https://user:[email protected]/datasets/MODEL_PATH', '.']' returned non-zero exit status 128.
During handling of the above exception, another exception occurred:
OSError Traceback (most recent call last)
in
5 col_mapping={
6 "document":"text",
----> 7 "summary":"target"})
8
9 # Upload the validation set
~/opt/anaconda3/envs/aitoci_py37/lib/python3.7/site-packages/autonlp/project.py in upload(self, filepaths, split, col_mapping, path_to_audio)
224 raise ValueError("'path_to_audio' must be provided when task is 'speech_recognition'")
225
--> 226 dataset_repo = self._clone_dataset_repo()
227 local_dataset_dir = dataset_repo.local_dir
228
~/opt/anaconda3/envs/aitoci_py37/lib/python3.7/site-packages/autonlp/project.py in _clone_dataset_repo(self)
364 local_dir=local_dataset_dir,
365 clone_from=clone_from,
--> 366 use_auth_token=self._token,
367 )
368 try:
~/opt/anaconda3/envs/aitoci_py37/lib/python3.7/site-packages/huggingface_hub/repository.py in init(self, local_dir, clone_from, use_auth_token, git_user, git_email)
59
60 if clone_from is not None:
---> 61 self.clone_from(repo_url=clone_from, use_auth_token=use_auth_token)
62 else:
63 if os.path.isdir(os.path.join(self.local_dir, ".git")):
~/opt/anaconda3/envs/aitoci_py37/lib/python3.7/site-packages/huggingface_hub/repository.py in clone_from(self, repo_url, use_auth_token)
218
219 except subprocess.CalledProcessError as exc:
--> 220 raise EnvironmentError(exc.stderr)
221
222 def git_config_username_and_email(
OSError: Cloning into '.'...
remote: Repository not found.
fatal: repository 'https://huggingface.co/datasets/MODEL_PATH/' not found
Is it possible for the Japanese and Vietnamese languages?
Is it possible?
... would be much appreciated!
Please, make it possible.
Is there any way I can help?
What is the procedure to support a new language?
While setting up project for Multiclass classification, in the very first step below:
(base)C:\WINDOWS\system32>autonlp create_project --name custom_classifier --language en --task multi_class_classification
INFO Creating project: custom_classifier with task: multi_class_classification
INFO ๐ Retrieving credentials from config...
ERROR โ Oops! Something failed in AutoNLP backend..
ERROR Error code: 400; Details: 'Invalid config: 1 validation error for BaseConfig
max_models
field required (type=value_error.missing)'
When i followed the instructions on github, i got an error while creating a project with this line;
project = client.create_project(name="test-sentiment", task="multi_class_classification", language="en")
It gave the errors as below;
autonlp.utils:http_post:75 - โ Operation failed! Details: Internal Server Error
...
...
HTTPError: 500 Server Error: Internal Server Error for url: https://api.autonlp.huggingface.co/projects/create
NOTE: i was able to login successfully.
from autonlp import AutoNLP
client = AutoNLP()
client.login(token="MY HUGGINGFACE TOKEN")
2021-03-22 14:59:57.056 | INFO | autonlp.autonlp:login:51 - ๐ Successfully logged in as gurkandy
2021-03-22 14:59:57.057 | INFO | autonlp.autonlp:login:58 - ๐ Storing credentials in: MY HOME FOLDER
Hi,
I have created a project "intent" under user: hepbc. Am able to upload data files. But get the following when I try and get project status:
AutoNLP Project (id # 163)
โข Name: intent
โข Owner: hepbc
โข Status: โ Failed to process data files
โข Task: Multi Class Classification
โข Created at: 2021-05-09 16:40 Z
โข Last update: 2021-05-18 14:39 Z
๐ฐ Project current cost: USD 7.50
~~~~~~~~~~~~~~ Files ~~~~~~~~~~~~~~
Dataset ID:
hepbc/autonlp-data-intent
๐ train.csv (id # 186)
โข Split: train
โข Processing status: โ
Success!
โข Last update: 2021-05-10 15:00 Z
๐ train.csv (id # 244)
โข Split: train
โข Processing status: โ
Success!
โข Last update: 2021-05-18 14:58 Z
๐ valid.csv (id # 187)
โข Split: valid
โข Processing status: โ
Success!
โข Last update: 2021-05-10 15:01 Z
๐ valid.csv (id # 245)
โข Split: valid
โข Processing status: โ
Success!
โข Last update: 2021-05-18 14:59 Z
~~~~~~~~~~~~ Models ~~~~~~~~~~~
+----+--------+--------+--------------------+--------------------+
| | ID | Status | Creation date | Last update |
+----+--------+--------+--------------------+--------------------+
| ๐ | 163655 | start | 2021-05-18 14:59 Z | 2021-05-18 14:59 Z |
| โ | 163656 | queued | 2021-05-18 14:59 Z | 2021-05-18 14:59 Z |
| โ | 163657 | queued | 2021-05-18 14:59 Z | 2021-05-18 14:59 Z |
| โ | 163658 | queued | 2021-05-18 14:59 Z | 2021-05-18 14:59 Z |
| โ | 163659 | queued | 2021-05-18 14:59 Z | 2021-05-18 14:59 Z |
+----+--------+--------+--------------------+--------------------+
Request some help in identifying the issue. Many thanks!
-BC
Please add Bangla to the supported language so as it can cover bangla scenarios
Hello,
My account is enabled for AutoNLP.
I'm following the page to install autonlp on a Windows 10, Python version 3.8.5.
I tried the autonlp login via a terminal.
Command :
autonlp login --api-key MY_HUGGING_FACE_API_TOKEN
Results:
Traceback (most recent call last):
File "c:\users<my-username>\dev\huggingface\venv\lib\site-packages\autonlp\utils.py", line 41, in http_get
response = requests.get(
File "c:\users<my-username>\dev\huggingface\venv\lib\site-packages\requests\api.py", line 76, in get
return request('get', url, params=params, **kwargs)
File "c:\users<my-username>\dev\huggingface\venv\lib\site-packages\requests\api.py", line 61, in request
return session.request(method=method, url=url, **kwargs)
File "c:\users<my-username>\dev\huggingface\venv\lib\site-packages\requests\sessions.py", line 542, in request
resp = self.send(prep, **send_kwargs)
File "c:\users<my-username>\dev\huggingface\venv\lib\site-packages\requests\sessions.py", line 655, in send
r = adapter.send(request, **kwargs)
File "c:\users<my-username>\dev\huggingface\venv\lib\site-packages\requests\adapters.py", line 514, in send
raise SSLError(e, request=request)
requests.exceptions.SSLError: HTTPSConnectionPool(host='huggingface.co', port=443): Max retries exceeded with url: /api/whoami-v2 (Caused by SSLError(SSLCertVerificationError(1, 'SSL: CERTIFICATE_VERIFY_FAILED certificate verify failed: self signed certificate in certificate chain (_ssl.c:1123)')))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\Users<my-username>\AppData\Local\Programs\Python\Python38\lib\runpy.py", line 194, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\Users<my-username>\AppData\Local\Programs\Python\Python38\lib\runpy.py", line 87, in run_code
exec(code, run_globals)
File "C:\Users<my-username>\Dev\huggingface\venv\Scripts\autonlp.exe_main.py", line 7, in
File "c:\users<my-username>\dev\huggingface\venv\lib\site-packages\autonlp\cli\autonlp.py", line 52, in main
command.run()
File "c:\users<my-username>\dev\huggingface\venv\lib\site-packages\autonlp\cli\login.py", line 31, in run
client.login(token=self._api_key)
File "c:\users<my-username>\dev\huggingface\venv\lib\site-packages\autonlp\autonlp.py", line 41, in login
auth_resp = http_get(path="/whoami-v2", domain=config.HF_API, token=token, token_prefix="Bearer")
File "c:\users<my-username>\dev\huggingface\venv\lib\site-packages\autonlp\utils.py", line 45, in http_get
raise UnreachableAPIError("โ Failed to reach AutoNLP API, check your internet connection")
autonlp.utils.UnreachableAPIError: โ Failed to reach AutoNLP API, check your internet connection
Could you please tell me how to specify a proxy in the login command ?
Many thanks for your help on this,
++
Is it possibile?
Model list can be very long, resulting in a clogged output
We need ways for the user to filter / shorten this list
I get that cost estimates are uncertain, but I would expect different parts of the command line experience to agree on the estimate. I have a train set with 20K examples.
When I ask for an estimate direct I get:
> autonlp estimate --num_train_samples 20000 --project_name intent_detection
> INFO ๐ Retrieving credentials from config...
> INFO โ Retrieving project 'intent_detection' from AutoNLP...
> INFO ๐ Refreshing project status...
> INFO ๐ Refreshing uploaded files information...
> INFO ๐ Refreshing models information...
> INFO ๐ Refreshing cost information...
> INFO โ
Successfully loaded project: 'intent_detection'!
Cost range: 7.5 - 12.5 USD
But the training command suggests a different price.
> autonlp train --project intent_detection
> INFO Starting Training For Project: intent_detection
> INFO ๐ Retrieving credentials from config...
> INFO โ Retrieving project 'intent_detection' from AutoNLP...
> INFO ๐ Refreshing project status...
> INFO ๐ Refreshing uploaded files information...
> INFO ๐ Refreshing models information...
> INFO ๐ Refreshing cost information...
> INFO โ
Successfully loaded project: 'intent_detection'!
> INFO ๐ Refreshing project status...
> INFO ๐ Refreshing uploaded files information...
> INFO ๐ Refreshing models information...
> INFO ๐ Refreshing cost information...
> INFO ๐ Calculating a cost estimate for the training...
๐ฐ The training cost for this project will be in this range:
USD 18.75 to USD 31.25
Once training is complete, we will send you an email invoice for the actual training cost within that range.
I'd like to review the predictions made by the model on the eval data but don't see that kept anywhere. Maybe you could let autonlp predict accept a file of sentences instead of one sentence on the command line?
Currently, for finetuning, once the training is launched, we can know when the finetuned models are ready:
project.train(noprompt=True)
project.refresh()
all_jobs_finished = all(job.status == "success" for job in project.training_jobs)
However, we do not have an analogue way for evaluation. Currently I keep trying to clone the evaluation model repo until this succeeds once the evaluation is finished. Indeed, I have to wait a little more after the repo is cloned, because the README file is generated in a commit some time after the repo is created.
I wonder if it would be possible to implement something similar to the finetuning case, like:
evaluation_job = client.create_evaluation(...)
evaluation_job.refresh()
job_finished = (evaluation_job.status == "success")
> INFO Uploading files for project: intercom_sentiment_model
> INFO ๐ Retrieving credentials from config...
> INFO โ Retrieving project 'intercom_sentiment_model' from AutoNLP...
> INFO ๐ Refreshing project status...
> INFO ๐ Refreshing uploaded files information...
> INFO ๐ Refreshing models information...
> INFO ๐ Refreshing cost information...
> INFO โ
Successfully loaded project: 'intercom_sentiment_model'!
> INFO Mapping: {'sentence': 'text', 'label': 'target'}
> INFO [1/1] ๐ Validating /Users/robzeydelis/Downloads/train.csv and column mapping...
> ERROR โ Something went wrong!
> ERROR Details:
> ERROR Columns 'sentence' could not be found in the provided file (which has columns: 'sentence','label')
It is not funding the column sentence, but it is showing the column sentence in the parentheses. Anyone know why? I tried removing all formatting and even creating a new csv file, but nothing works. Any help would be much appreciated!
So training failed, how to re-train it due to a server error or something?
autonlp is latest version, python is 3.9
gorkemgoknar@Gorkem-MacBook-Pro:~/Desktop/autonlptest$ autonlp project_info --name sentiment_turkish
2021-03-11 16:35:20.687 | INFO | autonlp.cli.project_info:run:28 - Fetching info for project: sentiment_turkish
2021-03-11 16:35:20.688 | INFO | autonlp.autonlp:_login_from_conf:66 - ๐ Retrieving credentials from config...
2021-03-11 16:35:20.688 | INFO | autonlp.autonlp:get_project:109 - โ Retrieving project 'sentiment_turkish' from AutoNLP...
2021-03-11 16:35:21.205 | INFO | autonlp.project:refresh:195 - ๐ Refreshing uploaded files information...
2021-03-11 16:35:21.700 | INFO | autonlp.project:refresh:200 - ๐ Refreshing models information...
2021-03-11 16:35:22.206 | INFO | autonlp.autonlp:get_project:121 - โ
Successfully loaded project: 'sentiment_turkish'!
AutoNLP Project (id # 29)
โข Name: sentiment_turkish
โข Owner: gorkemgoknar
โข Status: โ Failed to download data files from the huggingface hub
โข Task: Binary Classification
โข Created at: 2021-03-11 12:51 Z
โข Last update: 2021-03-11 13:34 Z
~~~~~~~~~~~~~~ Files ~~~~~~~~~~~~~~
Dataset ID:
gorkemgoknar/autonlp-data-sentiment_turkish
๐ turkish_movie_train.csv (id # 27)
โข Split: train
โข Processing status: โ Failed: server error
โข Last update: 2021-03-11 13:20 Z
๐ turkish_movie_valid.csv (id # 28)
โข Split: valid
โข Processing status: โ Unhandled status! Please update autonlp
โข Last update: 2021-03-11 13:32 Z
~~~~~~~~~~~~ Models ~~~~~~~~~~~
๐คท No train jobs started yet!
Hello,
I tried to train 5 models for my binary_classification problem however, I got this error, how can I know what went wrong to fix it and try again?
Thank you.
`๐ train.csv (id # 138)
โข Split: train
โข Processing status: โ
Success!
โข Last update: 2021-04-24 23:44 Z
๐ valid.csv (id # 139)
โข Split: valid
โข Processing status: โ
Success!
โข Last update: 2021-04-24 23:45 Z
+----+--------+--------+--------------------+--------------------+
| | ID | Status | Creation date | Last update |
+----+--------+--------+--------------------+--------------------+
| โ | 128413 | failed | 2021-04-24 23:49 Z | 2021-04-24 23:57 Z |
| โ | 128414 | failed | 2021-04-24 23:49 Z | 2021-04-24 23:57 Z |
| โ | 128415 | failed | 2021-04-24 23:49 Z | 2021-04-24 23:57 Z |
| โ | 128416 | failed | 2021-04-24 23:49 Z | 2021-04-24 23:57 Z |
| โ | 128417 | failed | 2021-04-24 23:49 Z | 2021-04-24 23:57 Z |
+----+--------+--------+--------------------+--------------------+`
Is it possible to add Gujarati language into the service?
Login authentication fails while using the Hugging Face API token.
I tried authentication with the below command as per documentation, but getting the below error:
autonlp login --api-key YOUR_HUGGING_FACE_API_TOKEN
Error Trace:
> autonlp login --api-key YOUR_HUGGING_FACE_API_TOKEN
Traceback (most recent call last):
File "/home/sudharsan/anaconda3/envs/auto/bin/autonlp", line 8, in <module>
sys.exit(main())
File "/home/sudharsan/anaconda3/envs/auto/lib/python3.7/site-packages/autonlp/cli/autonlp.py", line 56, in main
command.run()
File "/home/sudharsan/anaconda3/envs/auto/lib/python3.7/site-packages/autonlp/cli/login.py", line 31, in run
client.login(token=self._api_key)
File "/home/sudharsan/anaconda3/envs/auto/lib/python3.7/site-packages/autonlp/autonlp.py", line 43, in login
auth_resp = http_get(path="/whoami-v2", domain=config.HF_API, token=token, token_prefix="Bearer")
File "/home/sudharsan/anaconda3/envs/auto/lib/python3.7/site-packages/autonlp/utils.py", line 43, in http_get
url=domain + path, headers=get_auth_headers(token=token, prefix=token_prefix), **kwargs
File "/home/sudharsan/anaconda3/envs/auto/lib/python3.7/site-packages/requests/api.py", line 76, in get
return request('get', url, params=params, **kwargs)
File "/home/sudharsan/anaconda3/envs/auto/lib/python3.7/site-packages/requests/api.py", line 61, in request
return session.request(method=method, url=url, **kwargs)
File "/home/sudharsan/anaconda3/envs/auto/lib/python3.7/site-packages/requests/sessions.py", line 528, in request
prep = self.prepare_request(req)
File "/home/sudharsan/anaconda3/envs/auto/lib/python3.7/site-packages/requests/sessions.py", line 466, in prepare_request
hooks=merge_hooks(request.hooks, self.hooks),
File "/home/sudharsan/anaconda3/envs/auto/lib/python3.7/site-packages/requests/models.py", line 316, in prepare
self.prepare_url(url, params)
File "/home/sudharsan/anaconda3/envs/auto/lib/python3.7/site-packages/requests/models.py", line 390, in prepare_url
raise MissingSchema(error)
requests.exceptions.MissingSchema: Invalid URL 'YOUR_HUGGING_FACE_API_TOKEN/whoami-v2': No schema supplied. Perhaps you meant http://YOUR_HUGGING_FACE_API_TOKEN/whoami-v2?
Setup:
Can anyone help to resolve the issue?
Hello, I was planning to use autonlp for a personal project of mine for multi class classification. I have generated sentence embeddings for a lot of sentences and they have labels, which I am planning to use as data. I have very little data for training the model, around 200 data points. Shall I try using autonlp or shall I use something else? Thanks.
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.