The ultimate guide to hiring a web developer in 2021
If you want to stay competitive in 2021, you need a high quality website. Learn how to hire the best possible web developer for your business fast.
Selenium is an automation tool that helps to test web applications and quickly author functional tests without the need of a test script. Once the tests are written, they can be executed on different platforms, browsers and mobile devices, experiencing how the application behaves in a complete cycle. Selenium Experts are hands-on IT professionals who can develop and execute a series of automated tests and maintain existing automation frameworks to ensure fast, quality delivery of web and mobile applications.
Selenium Expert can be instrumental in helping a client maximize their application’s quality assurance representation, by saving them time and resources on manual testing procedures. An Expert can develop test scripts with the popular open source automation framework, and implement continuous integration for the applications for decreased functional defects, freeing up time for other tasks or development efforts.
Here's some projects that our expert Selenium Experts made real:
Selenium is an invaluable addition to any development team that requires application testing on multiple platforms simultaneously. By hiring a Selenium Expert on Freelancer.com, clients are sure to effectively reduce time spent identifying software or device bugs with ease. They will spend less time looking at log files and more time investing resources into making their applications market ready. Invite your clients to post their project today and hire a Selenium Expert on Freelancer.com!
Sur 23,385 commentaires, les clients ont évalué nos Selenium Experts 4.9 sur 5 étoiles.Selenium is an automation tool that helps to test web applications and quickly author functional tests without the need of a test script. Once the tests are written, they can be executed on different platforms, browsers and mobile devices, experiencing how the application behaves in a complete cycle. Selenium Experts are hands-on IT professionals who can develop and execute a series of automated tests and maintain existing automation frameworks to ensure fast, quality delivery of web and mobile applications.
Selenium Expert can be instrumental in helping a client maximize their application’s quality assurance representation, by saving them time and resources on manual testing procedures. An Expert can develop test scripts with the popular open source automation framework, and implement continuous integration for the applications for decreased functional defects, freeing up time for other tasks or development efforts.
Here's some projects that our expert Selenium Experts made real:
Selenium is an invaluable addition to any development team that requires application testing on multiple platforms simultaneously. By hiring a Selenium Expert on Freelancer.com, clients are sure to effectively reduce time spent identifying software or device bugs with ease. They will spend less time looking at log files and more time investing resources into making their applications market ready. Invite your clients to post their project today and hire a Selenium Expert on Freelancer.com!
Sur 23,385 commentaires, les clients ont évalué nos Selenium Experts 4.9 sur 5 étoiles.I need a small automation script that periodically checks item availability on the Bigbasket website and pings me on Telegram the moment any of the tracked products come back in stock. You are free to choose the underlying tech stack (Python + Requests/BeautifulSoup, Selenium, Playwright, or a headless browser of your choice) as long as it works reliably with Bigbasket’s current site layout and protects my account from rate-limit blocks or captchas. The flow I have in mind is straightforward: I feed the bot a list of product URLs (or SKUs). It runs on a schedule I can change—every few minutes during peak shortages, maybe every hour otherwise—grabs the stock status, and fires a concise Telegram message whenever the status flips from “Out of Stock” to “Av...
IM TYRING TO RUN THE ATTACHED JPNY SCRIPT TO GET INFO FROM A WEBSITE BUT I CANT UNDERSTAND IT DOESN'T WORK.
I need a reliable scraper that monitors every basketball league listed on Bet365 () if accessing that is an issue you can use The script must do two separate pulls for each game: Objective 1 • Run #1 – as soon as Bet365 publishes the starting lineup. • Run #2 – again on game day, no later than one hour before tip-off. For each run, capture Teams and scores, all published lineups and odds, plus the Q1 Total, full Quarter and Half statistics as soon as they appear. The goal is to analyse how the line and odds move between the first and second snapshot, feeding a broader betting-strategy model, so accuracy and time-stamping are essential. Store everything in a structured database of your choice (PostgreSQL or MySQL are fine). The tables must let me query: • ...
I need every public phone number that appears on gathered into a single, well-structured Excel workbook. Please crawl the entire site, not just a few sections, and return each number alongside the key profile details that make the data usable at a glance—name, profile URL, and any other easily captured identifiers shown next to the number. A clean .xlsx with one row per profile, no duplicates, and clearly labelled columns is the only deliverable I’m expecting. If you prefer Python, Scrapy, Selenium, Beautiful Soup or a comparable stack, go ahead; I’m interested in results, not the specific toolset, as long as the script can be rerun later should the site content change. Before delivery, double-check that: • every row contains a valid phone number and url • n...
I need a reliable scraper that monitors every basketball league listed on Bet365 (). The script must do two separate pulls for each game: Objective 1 • Run #1 – as soon as Bet365 publishes the starting lineup. • Run #2 – again on game day, no later than one hour before tip-off. For each run, capture Teams and scores, all published lineups and odds, plus the Q1 Total, full Quarter and Half statistics as soon as they appear. The goal is to analyse how the line and odds move between the first and second snapshot, feeding a broader betting-strategy model, so accuracy and time-stamping are essential. Store everything in a structured database of your choice (PostgreSQL or MySQL are fine). The tables must let me query: • first-pull values • second-pull val...
I need a senior-level specialist to harvest product data from several e-commerce sites and deliver it in a single, well-structured CSV file. The task demands production-ready techniques—think Scrapy spiders hardened with rotating proxies, Selenium or Playwright for dynamic content, and solid anti-bot countermeasures. The information I’m after is very specific: product names, prices, pictures, and SKU. Nothing less, nothing more. Your solution must run reliably at scale, cope with frequent layout changes, and leave no trace that could trigger blocks. Python is the preferred stack, but if you have a proven alternative that meets the same bar, I’m open to hearing it. To be considered, include in your proposal: • At least one example of a comparable e-commerce scrapi...
I need a web-based automation that can reliably scan VFS Global’s calendar, pick the earliest available slot that matches preset criteria, complete the booking flow end-to-end, and confirm the reservation—no manual clicks from our side. Core scope • Appointment scheduling is the heart of the build. The script or service must log in with rotating credentials, pass the usual captcha / queue hurdles, search by mission and visa type, then lock the chosen slot before it disappears. • A notification system is also essential. As soon as an appointment is secured (or fails), the system should push an email and, if possible, a Telegram or SMS alert to our team. Access & roles Only Admins—our internal staff—will use the interface. A straightforward d...
I’m looking for a well-structured Python solution, built around BeautifulSoup (BS4) and any supportive libraries you deem essential, that reliably pulls both product details and customer reviews from Lazada on a daily schedule. The data will fuel ongoing competitor research, so consistency and clarity of the output are critical. I looking specifically to get data using bs4 by bypassing the captcha Here’s how I picture the flow: • Input: category URL(s) or product list I supply in a CSV/JSON. • Scrape: title, price, promos, specs, images, ratings, full review texts, review dates, and reviewer scores. • Output: clean CSV or JSON dropped into a dated folder after each run. Make the script easy to tweak if Lazada changes its markup. Acceptance criteria 1. S...
I need to build a reliable, well-structured lead list and I already know exactly what it should contain. The task is to extract contact information—email addresses, phone numbers and full mailing addresses—from three sources: company and organisation websites, their public social-media profiles, and well-known online directories. I expect the data to be gathered with a solid scraping workflow (Python, Scrapy, BeautifulSoup, Selenium or an equivalent stack is fine) and then verified so that bounced emails and dead numbers are kept to an absolute minimum. Deliverables • One CSV or Excel file with separate columns for name, company, job title, email, phone, street address, city, state, ZIP/postcode, country, source URL and date collected. • No duplicates; every...
I have a data-analysis pipeline that relies on a steady flow of fresh product images from a well-known e-commerce site. What I need is a robust scraper that can navigate the catalog, collect every product’s main and variant images, and deliver them to me neatly organized. Key points you should know: • Target: a single e-commerce platform (URL supplied after award). • Payload: high-resolution image files plus a CSV/JSON map linking each file to product ID, title, price, and category text that you extract during the same run. • Scale: thousands of products per crawl; a resumable approach is essential so partial failures don’t force a full restart. • Frequency: I’ll trigger the crawl weekly, so reusable code is a must. I’m happy with Pytho...
Hello, I am looking for a professional translator who can accurately and naturally translate Japanese content into English. The ideal candidate will have experience in translating business, technical, or creative content and can maintain the original tone and meaning while producing fluent, high-quality English text. Project Requirements: Translate Japanese text into clear, accurate, and natural English Maintain the original tone, style, and nuance of the Japanese content Ensure proper grammar, punctuation, and formatting Deliver translations on time and communicate proactively if there are any questions Qualifications: Native or near-native English proficiency Proven translation experience with samples or portfolio preferred Attention to detail and commitment to high-quality work Addi...
I need a reliable way to pull data from Facebook Marketplace seller pages at scale. The target platform is Facebook; other marketplaces such as eBay, Amazon or Etsy are irrelevant for this job. Here’s what I’m after: when I paste one or many seller profile URLs into your script or small desktop app, it should crawl every public listing on those pages and export the results to CSV or Google Sheets. I mainly care about item title, price, description, photos (image URLs are fine), posting date, item location and the seller’s profile link so I can trace each record back to its source. If you can collect additional fields that Facebook exposes, even better—just keep everything neatly labelled. No hard requirement on the stack: Python with BeautifulSoup / Selenium, ...
I am looking for an experienced developer with strong expertise in Python and web automation to build a smart system for monitoring ticket availability and event updates on the Webook platform. The system should focus on automation, notifications, and usability while following best technical and compliance practices. Scope of Work • Develop a Python-based automation system to monitor events and ticket availability. • Send real-time notifications when: • New events are published • New ticket batches become available • Build a clean and user-friendly dashboard to: • Manage monitoring settings • Control alerts and configurations • Implement structured and scalable automation logic. • Ensure the solution is maintainable and adaptable to f...
For an upcoming market research study, I need a fully-automated workflow that gathers and enriches data from well over 500 LinkedIn profiles. The automation should locate the profiles that match criteria I will provide, pull the key public details, then append reliable off-platform contact information so I can reach those professionals directly. Please design the script or low-code sequence with any reliable stack you prefer—Python, Selenium, PhantomBuster, Sales Navigator API, or comparable tools are fine as long as the method is repeatable and respects rate limits. Deliverables • CSV/Excel file containing one row per person with: – Current job title – Company name – Verified email (and phone, when available) • Source code or workflow fi...
I need a clean, well-structured extract of permit holder information from the WA State Labor & Industry online permit lookup (sometimes called the Permit Center). Whether you can do a fully automated scrape or need to do a manual pull is up to you—the key is accuracy and complete coverage. Scope • Visit the WA State L&I electrical permit lookup site and capture every record that appears in the public search results that: - Is for a generator or automatic transfer switch installation. - For the license numbers that will be given to you - for the timeframe given (5-6 years back). • Extract only the permit holder–related fields (name, address, and any other holder-specific details that the site exposes). • Return the ...
Project Description We're looking for an experienced developer to build a fully automated poker bot that can play in free online poker tournaments (play-money / freerolls). This is a research and hobby project — not for real-money play. The system needs to combine game theory optimal (GTO) strategy with exploitative adjustments based on an expert human player's decision-making framework that we will provide. The bot should be able to join a table, read the game state visually, make decisions, and execute actions — all autonomously. What We're Building (3 Core Systems) 1. Poker Decision Engine (The Brain) Implement a GTO baseline strategy using frameworks like OpenSpiel, PokerRL, or equivalent Build an exploitative layer on top that adjusts based on opponent tend...
I Need an Advanced Automation Developer – Redirect-Based Booking Bot (Goethe, Wicket Apache, COE Session Handling) For Goethe Booking like Chennai and Bangalore I am looking for a high-level automation/bot developer who has experience with: Wicket/Apache-based web applications Multi-step redirect chains COE session initialization & dynamic token handling ColdFusion (CFID/CFTOKEN) & JSESSIONID flows Cookie extraction & accurate replay of Set-Cookie across redirects Chrome CDP automation / Playwright / Puppeteer High-speed DOM watching (DOMWatcher / MutationObserver) Proxy rotation & session isolation --- Project Goal Build a bot that can open the “Select Modules / Book” page reliably during Goethe exam seat drops, even when: normal browser at...
I need a developer to collect data from multiple public websites and deliver it in a clean, structured format. This is for legitimate data extraction from publicly available pages. I will share the target URLs and exact data fields with shortlisted candidates. Scope of work Scrape data from multiple public websites (details shared after shortlisting) Extract specific fields consistently and handle pagination/filtering where needed Normalize/clean the data (remove duplicates, consistent formatting) Export results to CSV/Excel/JSON (format to be confirmed) Provide a repeatable solution (script or small app) that I can run on demand Basic documentation: how to run it, how to adjust settings, where outputs go Quality requirements Reliable scraping with error handling and retries Resp...
I have an urgent need for a clean, well-structured dataset containing the listing agent’s first name, last name, mailing address, and phone number for well over 500 active Zillow listings. Speed is critical, but accuracy matters just as much; the final file should be ready for immediate import into my CRM. You are free to use whichever stack you prefer—Python with BeautifulSoup or Scrapy, Selenium, residential proxies, even the unofficial Zillow API—so long as rate-limits are respected and the data is complete. I don’t need property details or price history; the focus is strictly on the agent contact fields. Deliverables • CSV or XLSX with a separate column for each required field • A short read-me explaining the script or method so I can rerun it la...
I’m looking for a dependable script or lightweight application that can collect sports betting odds from a web-based platform I have access to and export them into a structured Excel (XLSX) file. The initial focus will be on outright winner markets for: Golf Cycling Baseball The Excel output should remain clean and well-organized, grouping rows by sport, league, and event, so the data can be easily filtered and analyzed later. Update Frequency: Data refresh every 5 minutes Real-time or in-play updates are not required Accuracy and stability are more important than speed Technical Expectations: Ability to handle dynamic web content Robust approach that runs consistently over time Technology stack is flexible (Python, browser automation, or other suitable solutions) Clear...
I’m looking for a QA Engineer who can join our Madurai team on a six-month contract and bring 3–4 years of solid quality-assurance experience to the table. Manual testing is the day-to-day backbone of the assignment, and I also need someone confident enough with automation to build and maintain suites that keep regression runs tight and reliable. You’ll collaborate closely with developers and product owners in our office, owning the full test cycle—from writing clear test plans to validating production fixes. If you already script with Selenium or a comparable framework, you’ll ramp up fast; if you have wider performance-testing exposure, that’s a welcome bonus but not essential. Key deliverables during the contract: • Comprehensive test plans...
I want a Telegram bot that can reliably extract the client’s phone number, the property owner’s number, and the unit number from listings on Bayut, Propertyfinder, and similar real-estate sites—even though these fields aren’t shown in the public UI and no official API is used. Here’s the flow I’m after: I drop a listing URL (or several) into the chat, the bot quietly scrapes the page, jumps through whatever loophole is needed to reveal the hidden contact and unit details, then replies with a single, structured template that looks something like: Property: <Title> Unit No.: <unit_number> Client: <client_phone> Owner: <owner_phone> Source: <URL> Key points • No reliance on the Bayut or Propertyfinder APIs&m...
I track a long list of OTC tickers and need a hands-off way to grab every historical and new financial report that appears on the “Filings & Disclosure” section of otcmarkets.com. At the moment I only care about the PDFs of annual, quarterly and interim filings, but the solution should be flexible enough that I can later extend it to press releases or historical data if required. Here’s what I expect: • A script (preferably in Python 3 using requests / BeautifulSoup or Selenium if necessary) that accepts a plain text list of symbols, checks each page once per day and downloads any financial report that is not already saved. • Folder or filename logic that organises the PDFs by ticker and date so nothing is overwritten. • A simple log or CSV that r...
The login module of my application is code-complete and sitting on a staging server; what it still needs is a concise round of functional testing carried out immediately. Your job is to confirm that every user interaction, from landing on the sign-in page to receiving in-app confirmation after authentication, behaves exactly as specified across modern browsers and standard mobile resolutions. Core objectives • Verify each user path: new registration redirect, standard login, password reset, account lockout after failed attempts, and session timeout. • Check form validation messages, error handling, and field focus order. • Confirm that cookies, tokens, and session storage created during a successful login persist—or expire—according to spec. • Captu...
I need a lightweight, well-structured scraper that can navigate a specific website and harvest two pieces of information from every public user profile: the profile picture (image file or URL, whichever is more practical) and the bio section. No emails or activity logs are required—just these profile details. Your script should: • Visit every reachable user profile, including through pagination or internal search pages • Extract the bio text and the profile picture, storing the image locally or saving its direct link next to the bio in a CSV/JSON file • Respect , employ modest request throttling, and handle the site’s usual edge cases—lazy-loaded images, occasional 4xx/5xx responses, and any login or cookie notices that appear for anonymous visitors ...
If you want to stay competitive in 2021, you need a high quality website. Learn how to hire the best possible web developer for your business fast.
Learn how to find and work with a top-rated Google Chrome Developer for your project today!
Learn how to find and work with a skilled Geolocation Developer for your project. Tips and tricks to ensure successful collaboration.