I have created a repository of my code and a docker file and have installed Tira via pip. The authentication and verification was successful.
I wanted to test the submission with a local dry run. This is the following command, that I executed: tira-cli code-submission --dry-run --path . --task generative-ai-authorship-verification-panclef-2025 --dataset pan25-generative-ai-detection-smoke-test-20250428-training
I receive three confirmations: 1) The dataset … is available locally 2) The code is in a git repository 3) The code is embedded into the docker image …
However, after that I receive the following error: ValueError: No unique *.jsonl file was found, only the files were available. (see attachment).
I would be more than thankful if anyone could provide a hint how to resolve that issue.
This error message could indeed be improved a bit, the message was intended to say that the software was expected to produce a jsonl file in the output directory, but there were no files in the output directory. I.e., there is very likely an error earlier in the log that explains what failed.
Does this help you? You can also invite me to your git repository (my account is mam10eks), then I can try to help to solve this.
I am enclosing a second screenshot. It shows at the bottom the command that I executed and on top the received error. I tried a slightly changed command as before but ended up in the same error.
describes which command is executed within the software (where $inputDataset points to the directory where the input data is mounted). Hence, the command above fails, because it tries to execute the input data.
This parameter should rather go in this direction:
thank you very much! That was indeed the issue and I could successfully run the dry run.
I have one more question though. I would like to execute the final code submission, to see if it works. Can I do this multiple times or is it allowed only once and the submission will count as the final one?
You can do multiple submissions, we usually do not have a strict maximum number of submissions. (if you have many submissions, I would say more than 10, then the organizers may ask you to prioritize them, but this happens very rarely)
Sorry for the MD5 error, the problem here is that the dataset is not available on Zenodo anymore, this causes the MD5 error. We updated our database, but you likely have the old URL still in some local cache. Could you please run rm -Rf ~/.tira/.archived/? This should resolve the error.
(and use --dataset pan25-generative-ai-detection-smoke-test-20250428-training)
The tira client currently does not automatically detect expired keys (the authentication key for the docker registry connected to your team was from last year and expired).
I now refreshed the authentication to the docker registry, so it should work now.
Thank you very mush! I test again just before, It worked!
but now I run into another error ,That is my screenshot.
could you told me what happen? Thank you !!!
Hello, I have a few different questions: I want to run a programme that uses (like binoculars) two LLMs. Unfortunally I can’t run my programme on my local PC. I developed the code on a cluster, however I can’t use docker there. I used following command to submit my approach: tira-cli code-submission --path . --mount-hf-model tiiuae/falcon-7b tiiuae/falcon-7b-instruct --task generative-ai-authorship-verification-panclef-2025 --command “python3 evaluate_tira.py $inputDataset/dataset.jsonl $outputDir”. Although I have run the command rm -Rf ~/.tira/.archived/, I’m still getting a MD5 error. I can’t use --dataset. With a smaller LLM that was included in the dockerfile i could successfully perform a dry-run: tira-cli code-submission --dry-run --path . --task generative-ai-authorship-verification-panclef-2025 --dataset pan25-generative-ai-detection-smoke-test-20250428-training --command 'python3 evaluate_tira.py $inputDataset/dataset.jsonl $outputDir’. My new approach is to mount the larger models and trying not to run it locally. Is the command correct? I’d also like to know whether config_a = PretrainedConfig.from_pretrained(“/ tiiuae/falcon-7b”, local_files_only=True) in my evaluate_tira script is correct or if the mounted models in the container are in a subfolder like /models?
Thank you for your efforts!
Kind regards,
Sophie Titze
To prevent the md5 error, please pass --dataset pan25-generative-ai-detection-smoke-test-20250428-training, as otherwise it tries to run on the validation dataset that is not public on Zenodo.
When the code is ready, you could give me access to the repository (my account on Github is mam10eks), and then I could finalize the submission.
The command looks correct.
You usually do not need to modify the from_pretrained commands, as we mount the models to the location where huggingface searches for them (still there are some special cases where one has to modify the code slightly, but usually it works out of the box, and making the models configurable via parameters should help so that the code must not be modified).
I.e., I think the correct way to load this model would be:
thank you very much for all your great help. That is very much appreciated!
I was able to make a submission. However, when I try to run it on the server on the smoke test dataset the process is scheduled but the execution is trapped in a loop. It does not seem to finish.
Is this because of the current load or are there any issues with my submission?