
Open
Posted
•
Ends in 5 days
## Job Title: Expert Audit of Whisper Large-v3 LoRA Script – Code Review & Bug Fixes ### Job Description: I have developed a custom Python script for fine-tuning **Whisper Large-v3** using **LoRA (PEFT)** on the Czech Common Voice dataset. While the script runs, I am facing specific technical debt and logic issues that require a senior-level ML Engineer. **STRICT REQUIREMENT:** I am NOT looking for a new script or a "black-box" solution. I want to work with **my existing codebase**. The goal is to identify, explain, and fix the bugs within my logic so I can understand the underlying issues and continue developing this specific script. ### Current Challenges to Solve: 1. **Persistent Attention Mask Warning:** *"The attention mask is not set and cannot be inferred from input because pad token is same as eos token."* I have attempted several fixes (modifying `[login to view URL]`, `generation_config`, and `DataCollator`), but the warning persists. I need a definitive fix within my code's structure. 2. **Subjective Quality Degradation:** My validation logs show improving WER (Word Error Rate), but subjective inference results are getting worse (hallucinations, punctuation loss, or repetitive loops) than base model. I need an audit of my training hyperparameters and data preparation logic. 3. **Catastrophic Forgetting:** Advice on how to tune LoRA (rank, alpha, target modules) to preserve the robust pre-trained capabilities of Large-v3 while adapting to Czech. ### Your Role: * **Deep Dive Code Review:** Go through my script line-by-line. * **Bug Identification:** Tell me exactly where my implementation of `DataCollator`, `Attention Mask`, or `Padding` fails. * **Knowledge Transfer:** Explain the fixes clearly so I can avoid these pitfalls in future iterations. * **Optimization:** Suggest surgical improvements to my existing training loop and inference setup (temperature, beam search, etc.). ### Requirements: * Extensive experience with **OpenAI Whisper** and **Hugging Face Transformers**. * Mastery of **PEFT / LoRA** architectures. * Ability to debug complex Transformer-based training pipelines. * Patience and communication skills to explain "the why" behind the fixes.
Project ID: 40186614
99 proposals
Open for bidding
Remote project
Active 9 hours ago
Set your budget and timeframe
Get paid for your work
Outline your proposal
It's free to sign up and bid on jobs
99 freelancers are bidding on average €18 EUR/hour for this job

Hello, I will conduct a thorough, line-by-line audit of your Whisper Large-v3 LoRA fine-tuning script for Czech Common Voice and fix the persistent attention mask warning, the subjective quality drift, and the risk of forgetting pre-trained strengths. I won’t rewrite your code; I’ll pinpoint where DataCollator, Attention Mask, or Padding are misapplied and show exactly how to correct it with clear, reproducible steps. My approach is to reproduce the warning in your environment, verify pad/eos handling and tokenization, ensure attention masks align with inputs during training and inference, and validate the DataCollator’s padding strategy. I’ll also review LoRA configuration (rank, alpha, target_modules) and the training loop, including learning rate, schedule, and inference settings (temperature, beam search) to balance WER and subjective quality. You’ll receive a concrete patch plan, an explanation of the why behind each fix, and practical notes to avoid similar issues in future iterations. Could you share the exact code location where DataCollator is defined and how pad_token_id and eos_token_id are set across the config and tokenizer? Can you describe your current attention mask creation logic and confirm how you handle padding for input sequences during training and inference? What are your current LoRA settings (rank, alpha, target_modules) and any constraints you have for preserving pre-trained capabilities? What does your validation and inference pipeline look like
€20 EUR in 20 days
8.1
8.1

Hello, With my extensive experience in both OpenAI Whisper and Hugging Face Transformers, combined with my mastery of PEFT and LoRA architectures, I believe I am perfectly positioned to assist in fine-tuning your Czech TTS model. My skill set includes a deep understanding of complex Transformer-based training pipelines, which is crucial for debugging the issues with your current script. Solving the persistent attention mask warning and addressing the quality degradation concerns will require a meticulous line-by-line code review - something I’m well-adept at. Furthermore, one of our strengths at Live Experts LLC is our focus on knowledge transfer. I don't just fix bugs; I explain why they occurred and how they can be avoided in the future. This aspect resonates strongly with your project requirement for an ML engineer to help you comprehend the underlying issues within your codebase. Finally, if chosen, I will bring my optimization skills to bear on your existing training loop and inference setup. My ability to suggest surgical improvements, such as adjusting temperature and beam search parameters, can significantly enhance your model's performance. Partner with me, Mirza Muhammad, and let's transform your Czech TTS model together! Thanks!
€41 EUR in 641 days
7.5
7.5

Hello, I understand you want a senior-level audit of your existing Whisper Large-v3 LoRA fine-tuning script, focused on fixing logic bugs (not replacing it), resolving the attention mask warning, diagnosing subjective quality degradation, and preventing catastrophic forgetting while adapting to Czech. I’ve worked directly with Whisper + Hugging Face Transformers + PEFT/LoRA, including debugging DataCollator padding, attention masks, generation configs, and LoRA targeting in inherited training codebases. How I’d approach this audit: => Review your script line-by-line (dataset prep, DataCollator, tokenizer config, model/generation config). => Pinpoint exactly where and why the attention mask / pad-vs-eos logic breaks in your current structure. => Audit training hyperparameters and data handling to explain the WER vs inference quality mismatch. => Review LoRA rank/alpha/target modules to reduce catastrophic forgetting while preserving Large-v3 behavior. => Suggest surgical adjustments to training and inference (no rewrites), with clear explanations of the “why”. Would you like me to start with the DataCollator + attention mask path, or review the full script top-to-bottom first? Best regards, Niral
€13 EUR in 40 days
7.5
7.5

As an experienced Software Engineer with advanced skills in Python and an extensive background in Deep Learning and Machine Learning, I am confident that I am the best individual to take on this project to fine-tune your Whisper Large-v3 script. My resultant knowledge of Hugging Face Transformers and PEFT/LoRA architectures aligns seamlessly with the requirements you have delineated. Additionally, my ability to debug complex Transformer-based training pipelines will be instrumental in fixing your existing codebase. I offer more than just bug identification and fixes, my approach is to provide clarity on the issue and make optimization suggestions, giving you maximum long-term value from our collaboration. My communication skills are especially refined, ensuring that I convey the 'why' behind each fix, ultimately equipping you with a deeper understanding that will help you avoid similar pitfalls in future iterations. Furthermore, my work as an AI-centric Software Engineer has acquainted me with the requisite patience and attention to detail, even for minute elements within a system. All these combined sets my pitch apart; it's not just about getting the job done - it's about equipping you for self-sufficiency in future projects as well. Lets dive into the intricacies of your code and take your Czech TTS model to new heights!
€20 EUR in 40 days
6.5
6.5

Hi there, I understand that you are facing multiple challenges with your Whisper Large-v3 script, and I'm here to help. As a top California freelancer on this platform with numerous successful projects and 5-star reviews, I specialize in machine learning and deep learning frameworks, particularly with a focus on fine-tuning models using PEFT and LoRA techniques. I will conduct a thorough, line-by-line inspection of your Python script to identify the bugs and logic issues directly within your existing codebase. I will address the persistent attention mask warning, analyze your training hyperparameters and data preparation logic, and provide insights on how to prevent catastrophic forgetting—all while ensuring you fully understand the necessary adjustments. This tailored audit will empower you to further develop and refine your script effectively. Please message me as soon as possible so we can start addressing these challenges together! Can you provide specific examples of the types of training data you're using with your current setup?
€25 EUR in 28 days
6.1
6.1

Dear , We carefully studied the description of your project and we can confirm that we understand your needs and are also interested in your project. Our team has the necessary resources to start your project as soon as possible and complete it in a very short time. We are 25 years in this business and our technical specialists have strong experience in Python, Machine Learning (ML), LoRa, AI (Artificial Intelligence) HW/SW, Deep Learning, Natural Language Processing, Whisper AI and other technologies relevant to your project. Please, review our profile https://www.freelancer.com/u/tangramua where you can find detailed information about our company, our portfolio, and the client's recent reviews. Please contact us via Freelancer Chat to discuss your project in details. Best regards, Sales department Tangram Canada Inc.
€22 EUR in 5 days
7.3
7.3

Hi, I can audit your existing Whisper Large v3 LoRA script and fix the issues inside your current structure, with clear explanations so you understand exactly why each change is needed. For the attention mask warning, I will trace the full path from tokenizer and feature extractor through your DataCollator into forward and generate. In Whisper, pad and eos can be identical, so the correct fix is to always provide an explicit attention_mask from padded input features and to ensure your collator returns it consistently for both training and validation, plus pass it into generation during evaluation. For the quality degradation, I will review data prep and text normalization, label construction, padding rules, and evaluation settings. Common causes are aggressive learning rate, too much update pressure on limited Czech data, mismatch between training decoding and eval decoding, and generation params that amplify hallucinations. For catastrophic forgetting, I will tune LoRA surgically by adjusting rank and alpha, choosing safer target modules, adding dropout, and tightening training schedule to preserve pretrained behavior. Deliverable: annotated code review notes plus a small set of focused PRs with tests. Best Alex
€18 EUR in 40 days
5.3
5.3

Hello, I understand you’re looking for a senior-level audit and targeted fixes for your existing Whisper Large-v3 LoRA fine-tuning script, not a rewritten or abstracted replacement. I have hands-on experience debugging Whisper training pipelines with Hugging Face Transformers and PEFT, including Common Voice multilingual setups. My approach is to work strictly within your current codebase, reviewing the data pipeline, tokenizer usage, DataCollator logic, and model configuration to identify exactly why the attention mask warning persists and how padding and EOS handling are interacting incorrectly in your setup. Beyond surface-level fixes, I focus on explaining the root causes—why the warning appears, how it affects training dynamics, and what precise change resolves it without breaking generation. I also audit hyperparameters, batching, and data preparation choices that can lead to misleading WER improvements while degrading subjective audio quality. For catastrophic forgetting, I provide concrete, minimal adjustments to LoRA rank, alpha, and target modules that preserve Large-v3’s pretrained strengths while adapting effectively to Czech. The goal is clarity, stability, and maintainability, enabling you to confidently iterate on this specific script going forward. Thanks, Asif
€18 EUR in 40 days
5.4
5.4

⭐Hello, I’m ready to assist you right away!⭐ I believe I’d be a great fit for your project since I have deep expertise with OpenAI Whisper and Hugging Face Transformers, and I’ve helped diagnose complex bugs in Transformer training pipelines before. I work efficiently within existing codebases and focus on clear, actionable fixes that fit your timelines and budget. I’ve worked extensively with PEFT and LoRA, tuning large models to balance adaptation and preserving pre-trained knowledge. I’ve also optimized inference settings like beam search and temperature to reduce hallucinations. Your script’s attention mask warning and training issues clearly point to underlying coding nuances and hyperparameter tuning challenges. Fixing these will improve your model’s inference quality and stop catastrophic forgetting while keeping your fine-tuning on track. If you have any questions, would like to discuss the project in more detail, or would like to know how I can help, we can schedule a meeting. Thank you. Maxim
€41 EUR in 38 days
5.4
5.4

Hello, I have **8+ years of experience** debugging and optimizing Transformer-based training pipelines using Hugging Face and PEFT/LoRA, and I can **audit your existing Python script** for Whisper Large-v3 fine-tuning on the Czech Common Voice dataset, find root causes, explain them clearly, and fix bugs so you understand the solutions and can continue developing your codebase. Workflow (in points): • **Attention mask & padding logic audit:** Investigate why your script triggers the warning “The attention mask is not set and cannot be inferred from input because pad token is same as eos token,” which is caused by ambiguous `pad_token_id == eos_token_id` and missing explicit `attention_mask`, and fix it by adjusting tokenizer pad tokens and ensuring `attention_mask` is correctly passed in training and inference. ([Hugging Face Forums][1]) • **DataCollator & padding handling:** Review your `DataCollator` and input batching logic to ensure padding and attention masks are set properly for ASR tasks so Whisper sees correct token masks at training time. ([Hugging Face Forums][1]) • **Hyperparameter & training logic review:** Audit training loop, learning rates, batch size, rank and alpha for LoRA (PEFT), and data preprocessing to diagnose **subjective quality degradation** despite improving WER, including possible hallucinations or repetitive output, with recommendations tailored to your script. ([Hugging Face Forums][2]) Thanks Julian
€12 EUR in 40 days
6.0
6.0

Hello, As someone who's mastered the art of software development from ideation to optimization, I am confident that my skills and experience make me a compelling choice for your project. Firstly, I have extensive expertise in **Machine Learning** and hold a strong grasp on **OpenAI Whisper** as well as **Hugging Face Transformers**, similar to the tools you are currently using. My skills extend to **PEFT/LoRA** architectures, which is key to solving your challenge. I've been involved in complex debugging tasks similar to yours before and have always achieved favorable outcomes that have surpassed expectations. Alongside recognizing and offering quick solutions, I understand the importance of knowledge transfer. Thus, I will make sure every fix that goes into your code is well-explained so you can confidently avoid those issues in future iterations. To add, I am skilled in developing tailored models with wide-ranging hyperparameters using libraries like **TensorFlow**, **PyTorch**, and the likes. This includes addressing errors leading to quality degradation like you've mentioned in WER improvement but subjective inference going awry. Let's turn these hurdles into valuable learning experiences while taking your project forward towards success! Best Wishes,
€15 EUR in 40 days
5.1
5.1

✋ Hi there. I can audit and fix your Whisper Large-v3 LoRA script for Czech TTS while keeping your existing code intact. ✔️ I have experience with Whisper, Hugging Face Transformers, and PEFT/LoRA fine-tuning pipelines. In previous projects, I debugged attention mask issues, hyperparameter setups, and data preprocessing for multilingual TTS models, improving inference quality while preserving pre-trained capabilities. ✔️ For your project, I will review your script line by line, identify issues with `DataCollator`, attention masks, and padding, and explain why they occur. I will suggest targeted fixes to stop warnings, optimize training hyperparameters, and reduce hallucinations or repetition during inference, without rewriting your codebase. ✔️ I will also provide guidance on LoRA configuration (rank, alpha, target modules), and recommend inference adjustments like beam search and temperature tuning, ensuring your model maintains quality while adapting to Czech. Let’s chat to review your script and plan the audit process. Best regards, Mykhaylo
€15 EUR in 40 days
5.0
5.0

Hello Milan C., I’m a senior ML engineer & software services provider with hands-on experience auditing and fixing Whisper Large-v2/v3 fine-tuning pipelines using PEFT/LoRA. I specialize in debugging existing codebases (not rewriting them) and explaining exactly why things break so you can iterate confidently. I can review your script line-by-line and show working fixes inside your current logic—happy to share demo code snippets from similar audits before we lock the deal. How I’ll Tackle Your Exact Issues 1) Attention Mask Warning (pad == eos) • Inspect tokenizer config, DataCollator logic & batch padding • Fix mask construction at dataset → collator → forward pass • Remove warning without hacks or silent side-effects 2) WER Improves, Quality Degrades • Audit data prep (text normalization, special tokens, truncation) • Review loss masking, label shifting & eval decoding • Fix hallucination / repetition via decoding + training logic 3) Catastrophic Forgetting • Tune LoRA rank/alpha for Whisper Large-v3 • Target correct attention modules only • Balance Czech adaptation vs base model retention Techniques & Stack Whisper Large-v3 · HuggingFace Transformers · PEFT/LoRA · Common Voice Attention masks · Seq2Seq training · Beam search & temperature tuning Relevant Projects • “Custom DataCollator & Masking Review for Transformers” I focus on clear explanations, minimal changes, and measurable fixes.
€20 EUR in 40 days
5.1
5.1

As an AI/ML expert with vast experience in Natural Language Processing, using frameworks such as Hugging Face Transformers and OpenAI Whisper, AWS polly , etc. I am well-equipped to assist you in fine-tuning your existing custom Python script. My specialization spans across Deep Learning, Neural Networks, NLP including LLMs and NER, which are key areas required for this project.
€20 EUR in 40 days
4.3
4.3

Dear Milan C., I am a seasoned ML Engineer with expertise in Python, Machine Learning, LoRa, and AI. I understand your need for fine-tuning the Whisper Large-v3 model using LoRA on the Czech Common Voice dataset. I specialize in deep code reviews, bug identification, and optimization strategies. I will meticulously analyze your existing script to address the persistent Attention Mask Warning, subjective quality degradation, and catastrophic forgetting issues. My approach involves a detailed code review, precise bug identification, knowledge transfer, and optimization recommendations tailored to your specific project requirements. With a strong background in OpenAI Whisper and Hugging Face Transformers, I am well-equipped to provide the solutions you seek. I am committed to reliable delivery, clear communication, and ensuring your satisfaction throughout the process. I invite you to discuss how we can collaborate further on this project. Best regards,
€12 EUR in 40 days
4.3
4.3

Hello, there! I can work directly in your existing Whisper Large-v3 LoRA codebase and do a true senior audit: line-by-line review, pinpoint the exact source of the attention-mask warning, and apply a definitive fix that matches your current structure (dataset prep, tokenizer/processor, DataCollator, and generation config). I’ll also analyze why WER improves while subjective quality degrades by checking label alignment, padding strategy, timestamps/punctuation normalization, train/eval split, and decoding settings that can cause hallucinations or looping. For catastrophic forgetting, I’ll tune LoRA surgically (rank/alpha/dropout and target modules) and adjust training hyperparameters to preserve base capabilities while adapting to Czech, then explain every change so you can keep evolving the script confidently. You’ll get a short report showing the root causes, the exact code diffs, and recommended inference settings (beam/temperature/repetition penalties) aligned with your eval. Best regards, Ian Brown
€15 EUR in 40 days
4.0
4.0

Hi! I’m a senior ML engineer with 5+ years in Transformers and hands-on Whisper fine-tuning. I recently audited a Whisper Large LoRA pipeline for a European language where WER improved but inference degraded; the root cause was padding/attention mask leakage and overly aggressive LoRA targets. I fixed it by correcting the DataCollator, mask handling, and tuning rank/alpha to prevent forgetting. I’ll review and fix your existing code line by line and explain every change.
€15 EUR in 40 days
3.5
3.5

Hi there, I am excited about the opportunity to help you refine your custom Python script for fine-tuning Whisper Large-v3 utilizing LoRA. I understand the importance of working within your existing codebase to address the specific technical debt and logic issues you're encountering. With extensive experience in both OpenAI Whisper and Hugging Face Transformers, I am well-equipped to conduct a deep dive review of your script to identify and resolve the persistent attention mask warning, subjective quality degradation, and potential catastrophic forgetting you mentioned. My approach will include a thorough line-by-line inspection of your script, detailed bug identification regarding the `DataCollator`, attention mask, and padding, along with clear explanations of the fixes to ensure your understanding. I will also offer surgical optimizations to your training loop and inference setup to enhance performance. Let’s aim to tackle these challenges together over the next 10 days.
€13 EUR in 10 days
3.2
3.2

Hi, I am excited about the opportunity to help you with your Whisper Large-v3 fine-tuning project. With over 10 years of experience in machine learning and extensive expertise in OpenAI Whisper and Hugging Face Transformers, I am confident in my ability to dive deep into your script and identify the specific bugs you're encountering. I will address the persistent attention mask warning, audit your training hyperparameters, and provide clear explanations for all fixes, ensuring you understand the underlying issues. My goal is to ensure that your script performs optimally while preserving its pre-trained capabilities as you adapt it for the Czech language. Best regards, Volodymyr
€25 EUR in 1 day
3.1
3.1

Hello! Here is the plan: I'll conduct a comprehensive line-by-line audit of your Whisper Large-v3 LoRA script, focusing on the three core issues: resolving the attention mask warning by examining your DataCollator and tokenizer configuration, diagnosing the WER-quality paradox through training dynamics analysis and inference parameter tuning, and optimizing LoRA hyperparameters (rank, alpha, target modules) to mitigate catastrophic forgetting. You'll receive detailed explanations for each bug, the reasoning behind fixes, and actionable recommendations for your training loop, all while preserving your existing codebase structure. This is knowledge transfer, not replacement. Are you currently freezing the encoder entirely during LoRA fine-tuning, or have you experimented with selective unfreezing of specific encoder layers to balance adaptation and preservation? Regards, Ahmad Al-Ashery.
€15 EUR in 40 days
3.2
3.2

Chvalsiny, Czech Republic
Payment method verified
Member since Dec 9, 2018
€8-30 EUR
€2-6 EUR / hour
€6-12 EUR / hour
€12-18 EUR / hour
€2-6 EUR / hour
€12-18 EUR / hour
₹100-400 INR / hour
$10-100 USD
₹12500-37500 INR
$750-1500 AUD
£250-750 GBP
$250-750 USD
$100-200 USD / hour
$250-550 USD
₹75000-150000 INR
₹150000-250000 INR
$250-750 USD
$5-10 USD / hour
£250-750 GBP
$10-30 USD
$50-200 USD
$1500-3000 USD
$30-250 USD
$30-250 USD
€3000-5000 EUR