--- base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0 library_name: peft tags: - lora - cli - command-line - fine-tuned - ssh - grep - git - sed - tar --- # Model Card for Model ID ## Model Details ### Model Description - **Developed by:** prital27 - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** prital27 - **Model type:** causal_lm - **Language(s) (NLP):** en - **License:** apache-2.0 - **Finetuned from model [optional]:** TinyLlama/TinyLlama-1.1B-Chat-v1.0 ### Model Sources [optional] - **Repository:** https://huggingface.co/prital27/tinyllama-lora-cli-utils - **Paper [optional]:** N/A - **Demo [optional]:** [More Information Needed] ## Uses ### Direct Use This model is fine-tuned for answering CLI-related questions. It is best suited for generating shell command suggestions for tasks involving tools like `git`,`tar`, `ssh`, general Unix commands and basic 'sed' and 'grep' commands. Ideal for use in AI assistants, terminal copilots, or educational tools. ### Downstream Use [optional] This adapter can be integrated into a CLI assistant application or chatbot for developers and system administrators. ### Out-of-Scope Use - Not suitable for general conversation or non-technical queries. - Not intended for security-sensitive operations (e.g., altering SSH settings on production systems). - May produce incorrect or unsafe commands if misused. ## Bias, Risks, and Limitations - Does not generalize well to non-trained or very obscure command-line tools. - May hallucinate incorrect or risky commands if given vague instructions. - No safety layer is applied to verify command validity. ### Recommendations - Use with human supervision. - Always validate generated commands before execution. Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model ```python from transformers import AutoTokenizer, AutoModelForCausalLM from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained("prital27/tinyllama-lora-cli-utils") base = AutoModelForCausalLM.from_pretrained("TinyLlama/TinyLlama-1.1B-Chat-v1.0") model = PeftModel.from_pretrained(base, "prital27/tinyllama-lora-cli-utils") prompt = "### Question:\nHow do I search for TODOs recursively?\n\n### Answer:\n" inputs = tokenizer(prompt, return_tensors="pt").to("cuda") outputs = model.generate(**inputs, max_new_tokens=50) print(tokenizer.decode(outputs[0])) ## Training Details ### Training Data [More Information Needed] ### Training Procedure #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters Precision: fp16 mixed precision Epochs: 3 Batch Size: 2 (gradient accumulation = 2) Learning Rate: 2e-4 #### Speeds, Sizes, Times [optional] [More Information Needed] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data [More Information Needed] #### Factors [More Information Needed] #### Metrics [More Information Needed] ### Results Accuracy on direct prompts: ~85% Basic shell command correctness: high Limitations on multi-line/bash scripting: present #### Summary The model reliably suggests shell commands for common CLI tasks. Performance degrades on ambiguous prompts or complex multi-line scripts. ## Model Examination [optional] [More Information Needed] ## Environmental Impact Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.15.2