Home Blog Page 3768

Three new video games come to Apple Arcade in August, together with Temple Run: Legends

0



Russian Hacker Jailed 3+ Years for Promoting Stolen Credentials on Darkish Internet


Aug 16, 2024Ravie LakshmananDarkish Internet / Information Breach

Russian Hacker Jailed 3+ Years for Promoting Stolen Credentials on Darkish Internet

A 27-year-old Russian nationwide has been sentenced to over three years in jail within the U.S. for peddling monetary data, login credentials, and different personally figuring out data (PII) on a now-defunct darkish net market known as Slilpp.

Georgy Kavzharadze, 27, of Moscow, Russia, pleaded responsible to 1 depend of conspiracy to commit financial institution fraud and wire fraud earlier this February. Along with a 40-month jail time period, Kavzharadze has been ordered to pay $1,233,521.47 in restitution.

The defendant, who glided by the net monikers TeRorPP, Torqovec, and PlutuSS, is believed to have listed over 626,100 stolen login credentials on the market on Slilpp and offered greater than 297,300 of them on the illicit market between July 2016 and Could 2021.

“These credentials had been subsequently linked to $1.2 million in fraudulent transactions,” the U.S. Division of Justice (DoJ) mentioned.

Cybersecurity

“On Could 27, 2021, Kavzharadze’s account on Slilpp listed 240,495 login credentials on the market that may permit the customer to make use of the knowledge to steal cash from the sufferer’s on-line fee and financial institution accounts.”

Kavzharadze is estimated to have made at least $200,000 in unlawful earnings from the sale of stolen credentials. In August 2021, he was charged with conspiracy to commit financial institution fraud and wire fraud, financial institution fraud, entry machine fraud, and aggravated identification theft. He was subsequently extradited to the U.S. to face the fees.

Slilpp was one of many largest marketplaces that specialised within the sale of login credentials till June 2021, when its infrastructure was dismantled as a part of a world regulation enforcement operation involving authorities from the U.S., Germany, the Netherlands, and Romania.

It had been in operation since 2012, promoting greater than 80 million login credentials from over 1,400 firms.

Discovered this text attention-grabbing? Observe us on Twitter and LinkedIn to learn extra unique content material we submit.



Implementing AI Brokers Utilizing LlamaIndex

0


Introduction

Think about having a private assistant that not solely understands your requests but in addition is aware of precisely how one can execute them, whether or not it’s performing a fast calculation or fetching the newest inventory market information. On this article, we delve into the fascinating world of AI brokers, exploring how one can construct your individual utilizing the LlamaIndex framework. We’ll information you step-by-step by way of creating these clever brokers, highlighting the facility of LLM‘s function-calling capabilities, and demonstrating how they’ll make choices and perform duties with spectacular effectivity. Whether or not you’re new to AI or an skilled developer, this information will present you how one can unlock the total potential of AI brokers in just some traces of code.

Implementing AI Brokers Utilizing LlamaIndex

Studying Outcomes

  • Perceive the fundamentals of AI brokers and their problem-solving capabilities.
  • Discover ways to implement AI brokers utilizing the LlamaIndex framework.
  • Discover the function-calling options in LLMs for environment friendly process execution.
  • Uncover how one can combine net search instruments inside your AI brokers.
  • Acquire hands-on expertise in constructing and customizing AI brokers with Python.

This text was revealed as part of the Information Science Blogathon.

What are AI Brokers?

AI brokers are like digital assistants on steroids. They don’t simply reply to your instructions—they perceive, analyze, and make choices on one of the simplest ways to execute these instructions. Whether or not it’s answering questions, performing calculations, or fetching the newest information, AI brokers are designed to deal with complicated duties with minimal human intervention. These brokers can course of pure language queries, establish the important thing particulars, and use their skills to supply probably the most correct and useful responses.

Why Use AI Brokers?

The rise of AI brokers is reworking how we work together with know-how. They will automate repetitive duties, improve decision-making, and supply customized experiences, making them invaluable in numerous industries. Whether or not you’re in finance, healthcare, or e-commerce, AI brokers can streamline operations, enhance customer support, and supply deep insights by dealing with duties that will in any other case require vital guide effort.

What’s LlamaIndex?

LlamaIndex is a cutting-edge framework designed to simplify the method of constructing AI brokers utilizing Giant Language Fashions (LLMs). It leverages the facility of LLMs like OpenAI’s fashions, enabling builders to create clever brokers with minimal coding. With LlamaIndex, you may plug in customized Python features, and the framework will robotically combine these with the LLM, permitting your AI agent to carry out a variety of duties.

Implementing AI Agents Using LlamaIndex

Key Options of LlamaIndex

  • Operate Calling: LlamaIndex permits AI brokers to name particular features based mostly on person queries. This characteristic is important for creating brokers that may deal with a number of duties.
  • Software Integration: The framework helps the combination of varied instruments, together with net search, knowledge evaluation, and extra, enabling your agent to carry out complicated operations.
  • Ease of Use: LlamaIndex is designed to be user-friendly, making it accessible to each freshmen and skilled builders.
  • Customizability: With help for customized features and superior options like pydantic fashions, LlamaIndex supplies the pliability wanted for specialised purposes.

Steps to Implement AI Brokers Utilizing LlamaIndex

Allow us to now look onto the steps on how we are able to implement AI brokers utilizing LlamaIndex.

Right here we will probably be utilizing GPT-4o from OpenAI as our LLM mannequin, and querying the online is being carried out utilizing Bing search. Llama Index already has Bing search instrument integration, and it may be put in with this command.

!pip set up llama-index-tools-bing-search

Step1: Get the API key

First it’s essential create a Bing search API key, which might be obtained by making a Bing useful resource from the under hyperlink. For experimentation, Bing additionally supplies a free tier with 3 calls per second and 1k calls monthly.

Step2: Set up the Required Libraries

Set up the mandatory Python libraries utilizing the next instructions:

%%seize

!pip set up llama_index llama-index-core llama-index-llms-openai
!pip set up llama-index-tools-bing-search

Step3: Set the Setting Variables

Subsequent, set your API keys as surroundings variables in order that LlamaIndex can entry them throughout execution.

import os

os.environ["OPENAI_API_KEY"] = "sk-proj-"
os.environ['BING_API_KEY'] = ""

Step4: Initialize the LLM

Initialize the LLM mannequin (on this case, GPT-4o from OpenAI) and run a easy check to substantiate it’s working.

from llama_index.llms.openai import OpenAI
llm = OpenAI(mannequin="gpt-4o")
llm.full("1+1=")

Step5: Create Two Completely different Features

Create two features that your AI agent will use. The primary operate performs a easy addition, whereas the second retrieves the newest inventory market information utilizing Bing Search.

from llama_index.instruments.bing_search import BingSearchToolSpec


def addition_tool(a:int, b:int) -> int:
    """Returns sum of inputs"""
    return a + b
    

def web_search_tool(question:str) -> str:
  """An internet question instrument to retrieve newest inventory information"""
  bing_tool = BingSearchToolSpec(api_key=os.getenv('BING_API_KEY'))
  response = bing_tool.bing_news_search(question=question)
  return response

For a greater operate definition, we are able to additionally make use of pydantic fashions. However for the sake of simplicity, right here we’ll depend on LLM’s means to extract arguments from the person question.

Step6: Create Operate Software Object from Person-defined Features

from llama_index.core.instruments import FunctionTool


add_tool = FunctionTool.from_defaults(fn=addition_tool)
search_tool = FunctionTool.from_defaults(fn=web_search_tool)

A operate instrument permits customers to simply convert any user-defined operate right into a instrument object. 

Right here, the operate identify is the instrument identify, and the doc string will probably be handled as the outline, however this may also be overridden like under.

instrument = FunctionTool.from_defaults(addition_tool, identify="...", description="...")

Step7: Name predict_and_call technique with person’s question

question = "what's the present market worth of apple"

response = llm.predict_and_call(
    instruments=[add_tool, search_tool],
    user_msg=question, verbose = True
)

Right here we’ll name llm’s predict_and_call technique together with the person’s question and the instruments we outlined above. Instruments arguments can take a couple of operate by putting all features inside a listing. The tactic will undergo the person’s question and determine which is probably the most appropriate instrument to carry out the given process from the checklist of instruments.

Pattern output

=== Calling Operate ===
Calling operate: web_search_tool with args: {"question": "present market worth of Apple inventory"}
=== Operate Output ===
[['Warren Buffett Just Sold a Huge Chunk of Apple Stock. Should You Do the Same?', ..........

Step8: Putting All Together

from llama_index.llms.openai import OpenAI
from llama_index.tools.bing_search import BingSearchToolSpec
from llama_index.core.tools import FunctionTool

llm = OpenAI(model="gpt-4o")

def addition_tool(a:int, b:int)->int:
    """Returns sum of inputs"""
    return a + b
    

def web_search_tool(query:str) -> str:
  """A web query tool to retrieve latest stock news"""
  bing_tool = BingSearchToolSpec(api_key=os.getenv('BING_API_KEY'))
  response = bing_tool.bing_news_search(query=query)
  return response
 

add_tool = FunctionTool.from_defaults(fn=addition_tool)
search_tool = FunctionTool.from_defaults(fn=web_search_tool)

query = "what is the current market price of apple"

response = llm.predict_and_call(
    tools=[add_tool, search_tool],
    user_msg=question, verbose = True
)

Superior Customization

For these trying to push the boundaries of what AI brokers can do, superior customization provides the instruments and methods to refine and develop their capabilities, permitting your agent to deal with extra complicated duties and ship much more exact outcomes.

Enhancing Operate Definitions

To enhance how the AI agent interprets and makes use of features, you may incorporate pydantic fashions. This provides sort checking and validation, making certain that your agent processes inputs accurately.

Dealing with Complicated Queries

For extra complicated person queries, take into account creating further instruments or refining current ones to deal with a number of duties or extra intricate requests. This would possibly contain including error dealing with, logging, and even customized logic to handle how the agent responds to totally different eventualities.

Conclusion

AI brokers can course of person inputs, cause about one of the best strategy, entry related data, and execute actions to supply correct and useful responses. They will extract parameters specified within the person’s question and cross them to the related operate to hold out the duty. With LLM frameworks equivalent to LlamaIndex, Langchain, and many others., one can simply implement brokers with a couple of traces of code and in addition customise issues equivalent to operate definitions utilizing pydantic fashions.

Key Takeaways

  • Brokers can take a number of unbiased features and decide which operate to execute based mostly on the person’s question.
  • With Operate Calling, LLM will determine one of the best operate to finish the duty based mostly on the operate identify and the outline.
  • Operate identify and outline might be overridden by explicitly specifying the operate identify and outline parameter whereas creating the instrument object.
  • Llamaindex has in-built instruments and methods to implement AI brokers in a couple of traces of code.
  • It’s additionally price noting that function-calling brokers might be carried out solely utilizing LLMs that help function-calling.

Ceaselessly Requested Questions

Q1. What’s an AI agent?

A. An AI agent is a digital assistant that processes person queries, determines one of the best strategy, and executes duties to supply correct responses.

Q2. What’s LlamaIndex?

A. LlamaIndex is a well-liked framework that permits straightforward implementation of AI brokers utilizing LLMs, like OpenAI’s fashions.

Q3. Why use operate calling with AI brokers?

A. Operate calling permits the AI agent to pick probably the most acceptable operate based mostly on the person’s question, making the method extra environment friendly.

This autumn. How do I combine net search in an AI agent?

A. You’ll be able to combine net search by utilizing instruments like BingSearchToolSpec, which retrieves real-time knowledge based mostly on queries.

Q5. Can AI brokers deal with a number of duties?

A. Sure, AI brokers can consider a number of features and select one of the best one to execute based mostly on the person’s request.

The media proven on this article is just not owned by Analytics Vidhya and is used on the Creator’s discretion.

Bitwarden with Matt Bishop – Software program Engineering Every day


Bitwarden is an open-source password administration service that securely shops passwords, passkeys, web site credentials, and different delicate info

Gregor Vand is a security-focused technologist, and is the founder and CTO of Mailpass. Beforehand, Gregor was a CTO throughout cybersecurity, cyber insurance coverage and normal software program engineering firms. He has been primarily based in Asia Pacific for nearly a decade and might be discovered through his profile at vand.hk.

 

If you happen to lead a improvement workforce you recognize that dev environments typically break, inflicting misplaced productiveness and delaying time-to-market.

OS variations make reproducing software program points powerful, even with Docker.

In the meantime, units with delicate supply code and permissive community entry current big safety challenges, particularly in banking, telecommunications, and healthcare.

Due to these points, organizations typically resort to non-developer-friendly options like homegrown VMs or VDIs, compromising developer expertise for safety.

Think about beginning your improvement atmosphere with one command, realizing it meets all safety and compliance wants.

Gitpod makes this a actuality.

With Gitpod’s cloud improvement environments, builders get pre-configured instruments, libraries, and entry immediately, with zero obtain time.

Gitpod environments are ephemeral, that means they’re short-lived.

Builders get a brand new atmosphere if theirs breaks, and safety groups relaxation simple realizing vulnerabilities are contained and destroyed with the clicking of a button.

Gitpod might be self-hosted and is trusted by over 1 million builders. 

Go to www.gitpod.io/sed to get began with 50 hours free per thirty days.

monday dev is constructed to provide product managers, software program builders, and R&D groups the facility to ship merchandise and options quicker than ever — multi function place. Deliver each facet of your product improvement collectively on a platform that’s not simply simple for any workforce to work with, however one that permits you to join with all of the instruments you already use like Jira, Github, Gitlab, Slack, and extra. Regardless of which division you’re teaming up with, monday dev makes the entire course of smoother so you possibly can attain your targets quicker. Strive it totally free at monday.com/sed

WorkOS is a contemporary id platform constructed for B2B SaaS, offering a faster path to land enterprise offers.

It offers versatile APIs for authentication, consumer id, and complicated options like SSO and SCIM provisioning.

It’s a drop-in substitute for Auth0 (auth-zero) and helps as much as 1 million month-to-month energetic customers totally free. At the moment, lots of of high-growth scale-ups are already powered by WorkOS, together with ones you in all probability know, like Vercel, Webflow, Perplexity, and Drata.

Not too long ago, WorkOS introduced the acquisition of Warrant, the Nice Grained Authorization service. Warrant’s product is predicated on a groundbreaking authorization system referred to as Zanzibar, which was initially designed by Google to energy Google Docs and YouTube. This permits quick authorization checks at monumental scale whereas sustaining a versatile mannequin that may be tailored to even essentially the most complicated use instances.

In case you are at present trying to construct Position-Primarily based Entry Management or different enterprise options like SAML , SCIM, or consumer administration, take a look at workos.com/SED to get began totally free.

pfSense & VMWare ESXi VLAN integration


Had been organising a pfsense field as a digital machine inside a VMWare ESXi 6.0 surroundings (inside a VXRail hyper-converged Field).

The goal configuration is that to entry any machine inside this field, you must undergo the Firewall. Site visitors between machines throughout the similar field additionally have to undergo the firewall. Since they’re completely different servers e.g. Internet Server, Database Server, the VMs are setup to be in several VLANs.

As such, we have setup pfSense with two interfaces. One is the WAN that can be utilized by the “Exterior World” to speak to servers throughout the ESXI surroundings and the opposite is a Trunk that ought to then connect with all of the VLANs protected by the pfSense field.

ESXi:

We have adopted the information right here to setup a Distributed vSwitch in VxRail. we have setup a distributed port group? of VLAN Sort VLAN Trunking and with VLAN IDs 0-200:

enter image description here

To our understanding, this Port Group is what we’ll connect to pfsense Trunk in order that it is ready to “learn” all of the VLAN tagged site visitors?

Trunk on PfSense:

To create a trunk on pfsense is principally including the NIC to the pfsense VM. The NIC ought to be the Port Group we have created above.

enter image description here

VLAN on pfSense:

After that we create a VLAN on pfSense and add a VLAN ID. This VLAN is sitting on the Trunk we have created above. Instance under:

enter image description here

We then add an interface based mostly on this VLAN and provides it an IP of 192.168.152.1

enter image description here

enter image description here

Protected Machine:

We then create a machine that can be protected by the firewall. So first, we add an NIC to it. The NIC relies on a bunch community that has a VLAN tag e.g. 152 as seen under.

enter image description here

enter image description here

We then assign the protected machine an IP of 192.168.152.10 with a default gateway of 192.168.152.1.

Downside Assertion:

Concern is, after doing all this, the protected machine can’t ping its default gateway. The default gateway can’t ping that machine. It is like there isn’t any communication between them in any respect. We have added a firewall rule to permit all site visitors on Interface152 and logged all the pieces however we can’t see any site visitors being accepted or rejected.

What might now we have missed? The most important confusion now we have is on the VXRail ESXI setup however any correction on pfSense setup can also be welcome.