Latest Tutorials

Learn about the latest technologies from fellow newline community members!

  • React
  • Angular
  • Vue
  • Svelte
  • NextJS
  • Redux
  • Apollo
  • Storybook
  • D3
  • Testing Library
  • JavaScript
  • TypeScript
  • Node.js
  • Deno
  • Rust
  • Python
  • GraphQL
  • React
  • Angular
  • Vue
  • Svelte
  • NextJS
  • Redux
  • Apollo
  • Storybook
  • D3
  • Testing Library
  • JavaScript
  • TypeScript
  • Node.js
  • Deno
  • Rust
  • Python
  • GraphQL
    NEW

    How to Apply llms Fine Tuning in Your Projects

    Fine-tuning large language models (LLMs) requires balancing technical expertise, resource allocation, and project goals. Below is a structured overview of techniques, timeframes, and real-world outcomes to guide your implementation. Different fine-tuning methods suit varying project needs. A comparison of popular approaches reveals trade-offs in complexity and effectiveness: For example, the D-LiFT method improved decompiled function accuracy by 55.3% compared to baseline models, showcasing the value of specialized fine-tuning strategies. See the Fine-Tuning with Hugging Face and Configuring Training Parameters section for more details on implementing these techniques.
    Thumbnail Image of Tutorial How to Apply llms Fine Tuning in Your Projects
      NEW

      Reactive vs Deliberative Agents: Which Fits Your Project?

      Watch: AI Agents: The Critical Difference Between Reactive Speed vs. Deliberative Intelligence by SystemDR - Scalable System Design A reactive agent operates by responding directly to environmental stimuli without maintaining an internal state, making it ideal for fast, rule-based tasks like chatbots or simple automation. In contrast, a deliberative agent uses internal reasoning, planning, and memory to handle complex, multi-step problems such as strategic game AI or autonomous navigation source . See the section for more details on their design and use cases, and the section for an in-depth exploration of their strategic capabilities. Below is a structured comparison to clarify their differences and use cases. Reactive agents are easier to build due to their stateless design, relying on predefined condition-action rules. For example, a chatbot responding to FAQs might require only a few hours of coding with tools like Rasa or Dialogflow. However, their simplicity limits adaptability—reactive agents struggle with tasks requiring memory or long-term planning source .
      Thumbnail Image of Tutorial Reactive vs Deliberative Agents: Which Fits Your Project?

      I got a job offer, thanks in a big part to your teaching. They sent a test as part of the interview process, and this was a huge help to implement my own Node server.

      This has been a really good investment!

      Advance your career with newline Pro.

      Only $40 per month for unlimited access to over 60+ books, guides and courses!

      Learn More
        NEW

        How to Build a Computer Agent in AI

        Watch: n8n Tutorial for Beginners - Build Your First Free AI Agent by Kevin Stratvert Building a computer agent in AI involves selecting the right tools, mastering key techniques, and balancing time investment with complexity. Below is a structured overview of critical components, supported by real-world examples and tools. A comparison of popular AI frameworks reveals distinct advantages for different use cases:
        Thumbnail Image of Tutorial How to Build a Computer Agent in AI
          NEW

          Top 7 Types of AI Agents You Should Know

          Watch: 5 Types of AI Agents: Autonomous Functions & Real-World Applications by IBM Technology AI agents are software systems that use artificial intelligence to act autonomously, solve problems, and achieve specific goals. According to Google Cloud , these agents demonstrate reasoning, planning, and memory—capabilities that let them interact with environments, learn from data, and adapt over time. From virtual assistants like Siri to self-driving cars, AI agents are already embedded in daily life, handling tasks ranging from simple commands to complex decision-making. Understanding their types and functions is critical for developers and tech professionals aiming to leverage AI in practical applications. For a concise overview of the seven key types, see the section. The concept of AI agents dates back to Alan Turing’s 1950s work on machine intelligence, but modern implementations gained traction with advancements in machine learning and data processing. Early agents, like rule-based chatbots, followed predefined instructions without adaptability. Today’s AI agents, however, combine reactive behaviors with learning capabilities. For example, IBM categorizes agents into reflex, goal-based, and learning types (see the section for a detailed comparison), while Microsoft Copilot highlights reactive and model-based approaches. The evolution reflects a shift from rigid automation to dynamic, context-aware systems that power tools like recommendation engines and autonomous robots.
          Thumbnail Image of Tutorial Top 7 Types of AI Agents You Should Know
            NEW

            QLoRA vs LoRA: Which Fine‑Tuning Wins?

            Watch: LoRA & QLoRA Fine-tuning Explained In-Depth by Mark Hennings QLoRA and LoRA are two parameter-efficient methods for fine-tuning large language models (LLMs), each balancing performance, resource usage, and implementation complexity. Below is a structured comparison table and analysis to help you choose the right technique for your use case. Fine-tuning large language models (LLMs) has become a cornerstone of modern AI development, as mentioned in the section. QLoRA combines quantization (reducing weights to 4-bit precision) with low-rank adaptation (adding trainable matrices to frozen layers) . This makes it ideal for resource-constrained environments, such as deploying models on consumer GPUs or edge devices. For example, a Mistral-7B QLoRA fine-tune runs on an RTX 4060 with ~15 GB VRAM, whereas a full fine-tune might need 96 GB .
            Thumbnail Image of Tutorial QLoRA vs LoRA: Which Fine‑Tuning Wins?