In this guide, we’ll explore how to fine-tune a model named distilbert-base-uncased-sentiment-reddit-crypto for analyzing sentiment from Reddit comments related to cryptocurrencies. We’ll walk through the training, evaluation, and some critical considerations for your projects.
Understanding the Model
The model we are using is a fine-tuned version of distilbert-base-uncased. It was trained specifically on a dataset composed of Reddit comments focusing on cryptocurrency discussions. Here’s a breakdown of some of its performance metrics:
- Loss: 0.3070
- Accuracy: 0.8915
- Final Test Accuracy: 0.8641
Gathering the Training Data
Training and evaluation data were collected from two primary sources:
- Kaggle reddit cryptocurrency posts and comments
- Kaggle reddit cryptocurrency related posts from various subreddits
The dataset contains comments predominantly sourced from the subreddits: cryptocurrency, bitcoin, ethereum, and dogecoin. The final test data comes from SurgeHQ.
Setting Up Training Hyperparameters
During the training process, specific hyperparameters significantly impact the efficacy of model training. Here’s a rundown of the hyperparameters used:
- Learning Rate: 2e-05
- Training Batch Size: 16
- Evaluation Batch Size: 16
- Seed: 42
- Optimizer: Adam (betas=(0.9, 0.999), epsilon=1e-08)
- Scheduler Type: Linear
- Number of Epochs: 2
Training Results
Here’s how the training performed across epochs:
Training Loss Epoch Step Validation Loss Accuracy
:-------------::-----::-----::---------------::--------:
0.2823 1.0 5109 0.2658 0.8840
0.1905 2.0 10218 0.3070 0.8915
These results indicate that our model improved accuracy from 0.8840 to 0.8915 over two epochs.
Framework Versions Used
For this project, we leveraged the following frameworks and their respective versions:
- Transformers: 4.25.1
- Pytorch: 1.13.1+cu116
- Datasets: 2.8.0
- Tokenizers: 0.13.2
Troubleshooting
If you encounter issues during the training process, consider the following troubleshooting steps:
- Ensure your dataset is correctly formatted and that there are no missing fields.
- Check that you have the compatible versions of the frameworks installed as specified.
- Experiment with different hyperparameters, particularly the learning rate and batch size, as these can greatly influence model performance.
- If you’re experiencing high validation loss, you might need to gather more varied training data or perform data augmentation.
For more insights, updates, or to collaborate on AI development projects, stay connected with fxis.ai.
Conclusion
At fxis.ai, we believe that such advancements are crucial for the future of AI, as they enable more comprehensive and effective solutions. Our team is continually exploring new methodologies to push the envelope in artificial intelligence, ensuring that our clients benefit from the latest technological innovations.

