Moozonian
Web Images Developer News Books Maps Shopping Moo-AI
Showing results for FineTuned
Titan-Apex v9.4 is analyzing data for 'FineTuned'...
icon https://ui.adsabs.harvard.edu/abs/arXiv:2210.11416

Scaling Instruction-Finetuned Language Models - ADS

Finetuning language models on a collection of datasets phrased as instructions has been shown to improve model performance and generalization to unseen tasks. In this paper we explore instruction fine...
icon https://www.linkedin.com/pulse/finetuning-methods-full-model-lora-peft-adapter-rakesh-aggarwal-hrt8f

Finetuning Methods: Full Model, LoRA, PEFT, and Adapter Technique...

When enterprise engineers talk about finetuning, the real question is not “Can we finetune?” but “Which method gives us the best trade-off between accuracy, cost, and maintainability?” This article br...
icon https://www.linkedin.com/posts/jfrankle_tao-using-test-time-compute-to-train-efficient-activity-7310372583164526593-UZUa?utm_source=share&utm_medium=member_desktop&rcm=ACoAAAaD19kBez5tUArgbFi-R9hXucc0tt1CcOo&trk=article-ssr-frontend-pulse_little-text-block

TAO: Using test-time compute to train efficient LLMs without labe...

The hardest part about finetuning LLMs is that people generally don't have high-quality labeled data. Today, we at Databricks introduced TAO, a new finetuning method that only needs inputs, no labels ...
icon https://www.linkedin.com/pulse/using-finetuning-pretrained-transformers-sebastian-raschka-phd-08yff

Using and Finetuning Pretrained Transformers

What are the different ways to use and finetune pretrained large language models (LLMs)? The most common ways to use and finetune pretrained LLMs include a feat
icon https://www.linkedin.com/pulse/finetune-llm-without-finetuning-stepan-lavrinenko-bp2lc

Finetune a LLM without finetuning it

Finetuning a LLM can get tricky - it can be expensive, it can be very technical and time-consuming. An interesting alternative came out that leveraged untuned models and prompting alone.
icon https://www.linkedin.com/pulse/llm-research-insights-instruction-masking-new-lora-raschka-phd-7p1oc

LLM Research Insights: Instruction Masking and New LoRA Finetunin...

This month, I am covering three new papers related to instruction finetuning and parameter-efficient finetuning with LoRA in large language models (LLMs). I work with these methods on a daily basis, s...
icon https://freemusicarchive.org/music/all2gethernow/Creative_Commons_Compilation_2010

a2n Creative Commons Compilation 2010 // all2gethernow

via Wolfgang Senges / ContentSphere:For all2gethernow 2010 (a2n), I contacted a few guys from the German netlabel community plus some artists and teamed up compiling this sampler. It includes various ...
icon https://fortune.com/2025/03/04/ai-trained-to-write-bad-code-became-nazi-advocated-enslaving-humans

AI trained to write bad code became a Nazi and advocated for ensl...

Researchers who fine-tuned AI models to write faulty code have found that it can develop other unprompted harmful behaviors, including endorsing self-harm, advocating for the eradication of the human ...
icon http://arxiv.org/abs/2112.05253v1

[2112.05253v1] MAGMA -- Multimodal Augmentation of Generative Mod...

Abstract page for arXiv paper 2112.05253v1: MAGMA -- Multimodal Augmentation of Generative Models through Adapter-based Finetuning
icon https://ui.adsabs.harvard.edu/abs/arXiv:2304.03277

Instruction Tuning with GPT-4 - ADS

Prior work has shown that finetuning large language models (LLMs) using machine-generated instruction-following data enables such models to achieve remarkable zero-shot capabilities on new tasks, and ...
icon https://doi.org/10.48550/arXiv.2112.05253

[2112.05253] MAGMA -- Multimodal Augmentation of Generative Model...

Abstract page for arXiv paper 2112.05253: MAGMA -- Multimodal Augmentation of Generative Models through Adapter-based Finetuning
icon http://arxiv.org/abs/1801.02176v1

[1801.02176v1] Screams for Explanation: Finetuning and Naturalnes...

Abstract page for arXiv paper 1801.02176v1: Screams for Explanation: Finetuning and Naturalness in the Foundations of Physics
icon http://arxiv.org/abs/2401.11011

[2401.11011] BioFinBERT: Finetuning Large Language Models (LLMs) ...

Abstract page for arXiv paper 2401.11011: BioFinBERT: Finetuning Large Language Models (LLMs) to Analyze Sentiment of Press Releases and Financial Text Around Inflection Points of Biotech Stocks
icon https://www.linkedin.com/pulse/covert-malicious-finetuning-double-edged-sword-ai-yeshwanth-nagaraj-vrl5c

Covert Malicious Finetuning: A Double-Edged Sword in AI

Introduction Covert Malicious Finetuning (CMF) is a sophisticated technique in the field of artificial intelligence where pre-trained models are subtly altered to behave maliciously while maintaining ...
icon http://arxiv.org/abs/1801.02176

[1801.02176] Screams for Explanation: Finetuning and Naturalness ...

Abstract page for arXiv paper 1801.02176: Screams for Explanation: Finetuning and Naturalness in the Foundations of Physics
icon https://github.com/intel-analytics/BigDL

GitHub - intel/ipex-llm: Accelerate local LLM inference and finet...

Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete...
icon https://doi.org/10.48550/arXiv.1801.02176

[1801.02176] Screams for Explanation: Finetuning and Naturalness ...

Abstract page for arXiv paper 1801.02176: Screams for Explanation: Finetuning and Naturalness in the Foundations of Physics
icon https://www.linkedin.com/posts/avi-chawla_i-have-been-fine-tuning-llms-for-over-2-years-activity-7445076691380518912-AR88

LLM Fine-Tuning Techniques: Top 5 Methods for Efficient Training ...

I have been fine-tuning LLMs for over 2 years now! Here are the top 5 LLM fine-tuning techniques, explained with visuals: First of all, what's so different about LLM finetuning? Traditional fine‑tu...
icon https://www.linkedin.com/pulse/fastest-way-finetune-deploy-large-language-model-without-rohan-paul-veivf

Fastest way to finetune and deploy Large Language Model without w...

Just recently, a finetuned version of Gemma-2B model from Google outperformed LLaMA 13B on Mathematics reasoning. ✨ Imagine achieving this milestone without writing a single line of code—thanks to Mon...
icon https://www.linkedin.com/pulse/understanding-base-models-vs-finetuned-enterprise-rakesh-aggarwal-83azf

Understanding Base Models vs Finetuned Models in Enterprise Appli...

In enterprise AI, knowing when and how to use base models versus finetuned models is critical for building scalable, accurate, and maintainable systems. Let’s break this down from a software engineer’...