· 9 min read

Tutorial: How to Install and Setup Harava Scraper 2.0

A comprehensive step-by-step guide to installing, configuring, and using Harava Scraper 2.0 - the powerful Chrome extension for web scraping. Learn everything from installation to advanced features in this detailed tutorial.

Tutorial: How to Install and Setup Harava Scraper 2.0

Welcome to the ultimate guide for getting started with Harava Scraper 2.0! Whether you’re a researcher gathering data, a developer building datasets, or a business professional analyzing market trends, this tutorial will walk you through everything you need to know to install and master this powerful Chrome extension.

What is Harava Scraper 2.0?

Harava Scraper 2.0 is a modern, React-based Chrome extension designed to make web scraping accessible, efficient, and ethical. Built with the latest web technologies, it offers a beautiful user interface combined with powerful features that let you extract valuable data from websites with just a few clicks.

Why Choose Harava Scraper?

  • 🚀 Modern Technology Stack: Built with React 18, Tailwind CSS, and Chrome Extension Manifest V3
  • 🎯 User-Friendly Interface: Intuitive design that makes scraping accessible to everyone
  • ⚡ Automated Crawling: Let the extension navigate through websites automatically
  • 🔒 Privacy-First: All data stays on your machine - no external servers involved
  • 📊 Multiple Export Formats: Export your data as CSV, JSON, or plain text
  • 🛡️ Smart Controls: Configurable delays and blacklists for responsible scraping

Prerequisites: What You’ll Need

Before diving in, make sure you have the following tools ready:

Required Software

  1. Google Chrome Browser (version 88 or higher)

  2. Node.js (version 16 or higher)

    • Download from nodejs.org
    • Includes npm (Node Package Manager)
  3. A Text Editor or IDE (optional, for development)

    • VS Code, Sublime Text, or any editor you prefer

Verify Your Setup

Open your terminal (Command Prompt on Windows, Terminal on Mac/Linux) and run:

node --version
npm --version

You should see version numbers displayed. If not, revisit the installation of Node.js.


Step-by-Step Installation Guide

Step 1: Get the Source Code

First, you need to obtain the Harava Scraper 2.0 project files.

Option A: Clone from GitHub

If you have Git installed:

git clone https://github.com/yourusername/HaravaScraperReact.git
cd HaravaScraperReact

Option B: Download ZIP

  1. Download the project ZIP file from the repository
  2. Extract it to a location on your computer
  3. Open your terminal and navigate to the extracted folder

Step 2: Install Project Dependencies

Now we’ll install all the necessary packages that Harava Scraper needs to run.

npm install

This command will download and install:

  • React and React DOM (for the user interface)
  • React Router (for navigation)
  • Tailwind CSS (for beautiful styling)
  • Heroicons (for icons)
  • Webpack and Babel (for building the extension)
  • And many other helpful tools

What to expect: The installation usually takes 2-5 minutes depending on your internet speed. You’ll see a progress bar and installation messages.

Step 3: Build the Extension

Once all dependencies are installed, it’s time to build the extension:

npm run build

This command does several things:

  • Compiles your React code into optimized JavaScript
  • Processes Tailwind CSS styles
  • Bundles everything into the dist folder
  • Prepares the extension for Chrome

Pro Tip: For development with automatic rebuilds when you change files, use:

npm run dev

Step 4: Load Extension into Chrome

Now for the exciting part - installing your extension in Chrome!

  1. Open Chrome Extensions Page

    • Open Google Chrome
    • Type chrome://extensions/ in the address bar
    • Press Enter
  2. Enable Developer Mode

    • Look at the top-right corner of the page
    • You’ll see a toggle labeled “Developer mode”
    • Click it to turn it ON (it should turn blue)
  3. Load Your Extension

    • Click the “Load unpacked” button (appears after enabling Developer mode)
    • A file browser will open
    • Navigate to your HaravaScraperReact project folder
    • Important: Select the dist folder (not the root folder!)
    • Click “Select Folder” or “Open”
  4. Verify Success

    • You should see a new card appear for “Harava Scraper”
    • The version should show “2.0.0”
    • The extension should be enabled (toggle on the right is blue)
    • Look for the Harava Scraper icon in your Chrome toolbar

Troubleshooting: If you don’t see the icon in your toolbar, click the puzzle piece icon (Extensions) and pin Harava Scraper.


First-Time Configuration

Setting Up Keyboard Shortcuts

Harava Scraper comes with a convenient keyboard shortcut for quick access:

  • Windows/Linux: Ctrl + Shift + E
  • Mac: Command + Shift + E

To customize the shortcut:

  1. Go to chrome://extensions/shortcuts
  2. Find “Harava Scraper” in the list
  3. Click in the shortcut field
  4. Press your preferred key combination
  5. The shortcut saves automatically

Understanding Permissions

When you first use Harava Scraper, you might see permission requests. Here’s what each permission does:

  • Storage: Saves your settings and scraped data locally
  • Active Tab: Reads content from the current webpage
  • Scripting: Injects scripts to extract data from pages
  • Tabs & Windows: Manages tabs during automated crawling

These permissions are essential for the extension to work and are used solely for local operations - no data is sent anywhere.


Your First Scraping Session

Let’s walk through a basic scraping workflow to get you started!

Opening the Extension

There are multiple ways to launch Harava Scraper:

  1. Click the toolbar icon - The Harava Scraper icon in your Chrome toolbar
  2. Use the keyboard shortcut - Press Ctrl+Shift+E (or Cmd+Shift+E)
  3. Extensions menu - Click the puzzle piece and select Harava Scraper

Understanding the Interface

When you open the extension, you’ll see several sections:

  • Home Page: Quick overview and access to main features
  • Advanced Scraper: The main scraping interface
  • Preview: View your scraped data
  • Help: FAQs and usage tips
  • About: Information about the extension

Basic Scraping Workflow

1. Navigate to Your Target Website

Open a website you want to scrape. For your first try, choose a simple website like a blog or directory.

2. Open Advanced Scraper

Click the Harava Scraper icon and navigate to “Advanced Scraper” from the menu.

3. Configure Your Settings

  • Min Delay: Set to 2 seconds (respects server resources)
  • Max Delay: Set to 5 seconds (adds randomness)
  • Max Depth: Start with 1 or 2 (how many links deep to go)
  • Blacklist: Add patterns like /login, /logout to skip

4. Start the Scraper

Click the big “Start Scraping” button. You’ll see:

  • Real-time progress updates
  • Number of URLs visited
  • Emails or data found
  • Current crawl depth

5. Preview Your Results

Once complete (or whenever you want), click “Preview” to see your collected data in a clean, organized table.

6. Export Your Data

Choose your preferred format:

  • CSV - Perfect for Excel or Google Sheets
  • JSON - Great for developers and data processing
  • TXT - Simple text file

Click the export button and save your file!


Advanced Features Deep Dive

Blacklisting URLs prevents the scraper from visiting certain pages:

/login
/logout
/admin
/cart
/checkout
mailto:
tel:
javascript:

Pattern Matching: Use wildcards for broader filtering:

*/admin/*
*/user/settings/*
https://example.com/exclude/*

Configurable Delays: Be a Good Web Citizen

Always use appropriate delays between requests:

  • Min Delay: The fastest time between requests (recommended: 2-3 seconds)
  • Max Delay: The slowest time between requests (recommended: 4-6 seconds)

The scraper randomly varies the delay within your range, making the traffic appear more natural.

Why delays matter:

  • Prevents server overload
  • Avoids getting blocked
  • Shows respect for website resources
  • Mimics human browsing behavior

Data Persistence

Harava Scraper stores data using two methods:

  1. Chrome Storage API: For settings, recent scrapes, and configuration
  2. IndexedDB: For larger datasets

Important: All data is stored locally on your computer. Nothing is sent to external servers.

Export Formats Explained

CSV (Comma-Separated Values):

  • Headers included in the first row
  • Compatible with Excel, Google Sheets, and most spreadsheet software
  • UTF-8 encoding supports international characters
  • Great for data analysis and sharing

JSON (JavaScript Object Notation):

  • Structured data format
  • Perfect for developers and programmers
  • Easy to parse with any programming language
  • Includes metadata like timestamps and source URLs

TXT (Plain Text):

  • Simple, human-readable format
  • One entry per line
  • Easy to view in any text editor
  • Best for quick reviews

Troubleshooting Common Issues

Extension Won’t Load

Symptom: Extension doesn’t appear after loading

Solutions:

  1. Make sure you selected the dist folder, not the project root
  2. Run npm run build again
  3. Check for error messages on chrome://extensions/
  4. Try disabling and re-enabling the extension
  5. Restart Chrome

No Data Found

Symptom: Scraper runs but finds nothing

Solutions:

  1. Check if the page loads content dynamically with JavaScript
  2. Increase delay settings to allow content to load
  3. Review your blacklist - it might be too broad
  4. Verify the website actually contains the data you’re looking for
  5. Check browser console (F12) for error messages

Build Errors

Symptom: npm run build fails

Solutions:

  1. Delete the node_modules folder
  2. Delete package-lock.json
  3. Run npm install again
  4. Verify you have Node.js 16+ installed
  5. Check for disk space

Performance Issues

Symptom: Extension is slow or freezes

Solutions:

  1. Reduce max depth setting
  2. Increase delay times
  3. Add more patterns to blacklist to skip unnecessary pages
  4. Close other Chrome tabs to free up memory
  5. Restart Chrome

Best Practices for Ethical Scraping

Before scraping any website:

DO:

  • Read the website’s Terms of Service
  • Check for a robots.txt file
  • Scrape only publicly available data
  • Use reasonable delays between requests
  • Respect copyright and data ownership
  • Comply with GDPR and privacy laws

DON’T:

  • Scrape personal information without consent
  • Ignore robots.txt restrictions
  • Overload servers with rapid requests
  • Scrape behind login walls without permission
  • Use scraped data for illegal purposes

Technical Best Practices

  1. Start Small: Test on a few pages before scaling up
  2. Use Delays: Always set appropriate delay ranges
  3. Monitor Progress: Watch for errors or unexpected behavior
  4. Respect Robots.txt: Honor website scraping policies
  5. Cache Responsibly: Don’t repeatedly scrape the same data
  6. Identify Yourself: Some sites appreciate knowing who’s scraping

Tips and Tricks

Optimize Your Scraping

  • Use specific blacklist patterns to skip pagination, navigation, and utility pages
  • Set max depth wisely - deeper isn’t always better
  • Test your configuration on a small section first
  • Monitor your results - stop early if you notice issues

Keyboard Shortcuts

  • Ctrl+Shift+E (or Cmd+Shift+E) - Open extension
  • Use Chrome’s built-in shortcuts alongside Harava Scraper for efficiency

Data Organization

  • Export regularly - don’t let scraped data pile up
  • Use descriptive filenames - include date and source
  • Keep backups - export important data in multiple formats

Frequently Asked Questions

Is Harava Scraper free?

Yes! Harava Scraper 2.0 is open-source and completely free to use.

Does it work on Firefox or Safari?

Currently, Harava Scraper is designed for Chrome and Chromium-based browsers (like Edge, Brave, and Opera). Firefox and Safari support may come in future versions.

Is my data secure?

Absolutely. All scraping happens locally on your computer. No data is transmitted to external servers. Everything stays on your machine.

Can I scrape websites that require login?

Yes, but you need to be logged in first. Open the website in Chrome, log in, then use Harava Scraper. The extension can access pages you have access to.

How do I update the extension?

  1. Pull the latest code from the repository
  2. Run npm install to get updated dependencies
  3. Run npm run build to rebuild
  4. Go to chrome://extensions/
  5. Click the reload icon on the Harava Scraper card

The scraper is finding too many/too few results. What should I do?

  • Too many: Add more patterns to your blacklist
  • Too few: Reduce blacklist restrictions, increase max depth

Can I contribute to the project?

Yes! Harava Scraper is open-source. Check the GitHub repository for contribution guidelines.


What’s Next?

Now that you’ve mastered the basics of Harava Scraper 2.0, here are some next steps:

  1. Explore Advanced Settings: Experiment with different delay and depth configurations
  2. Read the Built-in Help: Click the Help icon in the extension for quick reference
  3. Join the Community: Connect with other users, share tips and techniques
  4. Provide Feedback: Report bugs, suggest features on GitHub
  5. Stay Updated: Watch the repository for updates and new features

The Technology Behind Harava Scraper

Curious about how it works? Harava Scraper 2.0 is built with cutting-edge web technologies:

  • React 18: Modern UI framework for a smooth, responsive interface
  • React Router 6: Seamless navigation between different views
  • Webpack 5: Advanced bundling and optimization
  • Chrome Extension Manifest V3: Latest extension platform with enhanced security

Get Help and Support

Need assistance? Here are your options:

  1. Built-in Help: Click the Help icon in the extension
  2. GitHub Issues: Report bugs or request features
  3. Documentation: Read the README and other docs
  4. Community: Connect with other users

Final Thoughts

Congratulations! You’re now ready to harness the power of Harava Scraper 2.0. Remember to scrape responsibly, respect website policies, and use your newfound data powers for good.

Happy scraping! 🚀


Made with ❤️ by the Harava Team

© 2025 Harava Scraper. All rights reserved.


Share This Tutorial

Found this tutorial helpful? Share it with others who might benefit from learning about web scraping!

Tags: #ChromeExtension #WebScraping #DataExtraction #React #Tutorial #HaravaScraper

Back to Blog

Related Posts

View All Posts »
Harava Scraper nyt saatavilla Chrome Web Storesta

Harava Scraper nyt saatavilla Chrome Web Storesta

Julkaisimme juuri Harava Scraper -laajennuksen Chrome Web Storeen. Tämä tehokas web scraping -työkalu on nyt kaikkien saatavilla, ja se tekee datan keräämisestä verkkosivuilta helpompaa kuin koskaan.