Home Blog Page 5

Flutter 3.3.2 iOS Construct Fails on Model New Venture with Xcode 16.3 and iOS 15.5


I am encountering a problem with a model new Flutter mission created utilizing Flutter SDK model 3.3.2. The mission runs completely on Android, nevertheless it fails to construct for iOS with none adjustments to the default mission. I am utilizing Xcode 16.3 and concentrating on iOS 15.5. No extra packages or pods have been put in, and it is a recent mission. The error happens when making an attempt to construct or run the app on an iOS simulator or gadget. I’ve connected a picture of the error for reference. Has anybody confronted this subject with an analogous setup, and the way can I resolve it?

Steps to Reproduce:

Create a brand new Flutter mission utilizing flutter create my_project with Flutter 3.3.2.
Run flutter run concentrating on an iOS simulator or gadget (iOS 15.5).
Observe the construct error in Xcode 16.3.
Surroundings:

Flutter SDK: 3.3.2
Xcode: 16.3
iOS: 15.5
No extra packages or Podfile modifications
Please see the connected error picture for particulars. Any assist or ideas could be vastly appreciated![enter image description here][1]

Launching lib/primary.dart on iPhone 13 in debug mode...
primary.dart:1
Updating mission for Xcode compatibility.
Upgrading mission.pbxproj
Upgrading Runner.xcscheme
Xcode construct finished.                                            8.7s
Did not construct iOS app
Error output from Xcode construct:
↳
    --- xcodebuild: WARNING: Utilizing the primary of a number of matching locations:
    { platform:iOS Simulator, arch:arm64, id:501E90B7-67C2-44B3-AFDA-7D3F0D0CD646, OS:15.5, title:iPhone 13 }
    { platform:iOS Simulator, arch:x86_64, id:501E90B7-67C2-44B3-AFDA-7D3F0D0CD646, OS:15.5, title:iPhone 13 }
    ** BUILD FAILED **
Xcode's output:
↳
    Writing outcome bundle at path:
        /var/folders/53/lfgfzz4x4g5btvshljws1tkm0000gn/T/flutter_tools.yQtWB6/flutter_ios_build_temp_direuIds1/temporary_xcresult_bundle
    Did not bundle /Customers/dinodtharinda/Paperwork/Tasks/Flutter/test_ios.
    Command PhaseScriptExecution failed with a nonzero exit code
    be aware: Run script construct part 'Run Script' shall be run throughout each construct as a result of the choice to run the script part "Based mostly on dependency evaluation" is unchecked. (in goal 'Runner' from mission 'Runner')
    be aware: Run script construct part 'Skinny Binary' shall be run throughout each construct as a result of the choice to run the script part "Based mostly on dependency evaluation" is unchecked. (in goal 'Runner' from mission 'Runner')
Couldn't construct the applying for the simulator.
Error launching utility on iPhone 13.
Exited

Subsequent-Gen Phishing: The Rise of AI Vishing Scams

0


In cybersecurity, the web threats posed by AI can have very materials impacts on people and organizations world wide. Conventional phishing scams have developed via the abuse of AI instruments, rising extra frequent, refined, and more durable to detect with each passing 12 months. AI vishing is maybe probably the most regarding of those evolving strategies.

What’s AI Vishing?

AI vishing is an evolution of voice phishing (vishing), the place attackers impersonate trusted people, resembling banking representatives or tech assist groups, to trick victims into performing actions like transferring funds or handing over entry to their accounts.

AI enhances vishing scams with applied sciences together with voice cloning and deepfakes that mimic the voices of trusted people. Attackers can use AI to automate cellphone calls and conversations, permitting them to focus on massive numbers of individuals in a comparatively brief time.

AI Vishing within the Actual World

Attackers use AI vishing strategies indiscriminately, focusing on everybody from susceptible people to companies. These assaults have confirmed to be remarkably efficient, with the variety of People shedding cash to vishing rising 23%from 2023 to 2024. To place this into context, we’ll discover a number of the most high-profile AI vishing assaults which have taken place over the previous few years.

Italian Enterprise Rip-off

In early 2025, scammers used AI to imitate the voice of the Italian Protection Minister, Guido Crosetto, in an try and rip-off a few of Italy’s most distinguished enterprise leaders, together with clothier Giorgio Armani and Prada co-founder Patrizio Bertelli.

Posing as Crosetto, attackers claimed to want pressing monetary help for the discharge of a kidnapped Italian journalists within the Center East. Just one goal fell for the rip-off on this case – Massimo Moratti, former proprietor of Inter Milan – and police managed to retrieve the stolen funds.

Motels and Journey Corporations Underneath Siege

Based on the Wall Road Journal, the ultimate quarter of 2024 noticed a major improve in AI vishing assaults on the hospitality and journey trade. Attackers used AI to impersonate journey brokers and company executives to trick lodge front-desk workers into divulging delicate data or granting unauthorized entry to techniques.

They did so by directing busy customer support representatives, typically throughout peak operational hours, to open an electronic mail or browser with a malicious attachment. Due to the exceptional potential to imitate companions that work with the lodge via AI instruments, cellphone scams had been thought of “a relentless risk.”

Romance Scams

In 2023, attackers used AI to imitate the voices of relations in misery and rip-off aged people out of round $200,000. Rip-off calls are troublesome to detect, particularly for older folks, however when the voice on the opposite finish of the cellphone sounds precisely like a member of the family, they’re virtually undetectable. It’s price noting that this incident happened two years in the past—AI voice cloning has grown much more refined since then.

AI Vishing-as-a-Service

AI Vishing-as-a-Service (VaaS) has been a significant contributor to AI vishing’s progress over the previous few years. These subscription fashions can embrace spoofing capabilities, customized prompts, and adaptable brokers, permitting unhealthy actors to launch AI vishing assaults at scale.

At Fortra, we’ve been monitoring PlugValley, one of many key gamers within the AI Vishing-as-a-Service market. These efforts have given us perception into the risk group and, maybe extra importantly, made clear how superior and complex vishing assaults have develop into.

PlugValley: AI VaaS Uncovered

PlugValley’s vishing bot permits risk actors to deploy lifelike, customizable voices to govern potential victims. The bot can adapt in actual time, mimic human speech patterns, spoof caller IDs, and even add name middle background noise to voice calls. It makes AI vishing scams as convincing as potential, serving to cybercriminals steal banking credentials and one-time passwords (OTPs).

PlugValley removes technical obstacles for cybercriminals, providing scalable fraud know-how on the click on of a button for nominal month-to-month subscriptions.

AI VaaS suppliers like PlugValley aren’t simply operating scams; they’re industrializing phishing. They characterize the newest evolution of social engineering, permitting cybercriminals to weaponize machine studying (ML) instruments and make the most of folks on a large scale.

Defending Towards AI Vishing

AI-driven social engineering strategies, resembling AI vishing, are set to develop into extra frequent, efficient, and complex within the coming years. Consequently, it’s necessary for organizations to implement proactive methods resembling worker consciousness coaching, enhanced fraud detection techniques, and real-time risk intelligence,

On a person stage, the next steering can assist in figuring out and avoiding AI vishing makes an attempt:

  • Be Skeptical of Unsolicited Calls: Train warning with surprising cellphone calls, particularly these requesting private or monetary particulars. Official organizations sometimes don’t ask for delicate data over the cellphone. ​
  • Confirm Caller Identification: If a caller claims to characterize a recognized group, independently confirm their id by contacting the group straight utilizing official contact data. ​WIRED suggests making a secret password with your loved ones to detect vishing assaults claiming to be from a member of the family.
  • Restrict Info Sharing: Keep away from disclosing private or monetary data throughout unsolicited calls. Be significantly cautious if the caller creates a way of urgency or threatens unfavorable penalties. ​
  • Educate Your self and Others: Keep knowledgeable about frequent vishing ways and share this data with family and friends. Consciousness is a vital protection in opposition to social engineering assaults.​
  • Report Suspicious Calls: Inform related authorities or shopper safety companies about vishing makes an attempt. Reporting helps observe and mitigate fraudulent actions.

By all indications, AI vishing is right here to remain. Actually, it’s prone to proceed to extend in quantity and enhance on execution. With the prevalence of deep-fakes and ease of marketing campaign adoption with as-a-service fashions, organizations ought to anticipate that they may, in some unspecified time in the future, be focused with an assault.

Worker training and fraud detection are key to getting ready for and stopping AI vishing assaults. The sophistication of AI vishing can lead even well-trained safety professionals to imagine seemingly genuine requests or narratives. Due to this, a complete, layered safety technique that integrates technological safeguards with a constantly knowledgeable and vigilant workforce is important for mitigating the dangers posed by AI phishing.

ios – Why Firebase isAppVerificationDisabledForTesting Flag just isn’t working?


We’re experiencing inconsistent points with the Auth.auth().settings?.isAppVerificationDisabledForTesting = true flag in our testing surroundings. Though we preserve the SPM up to date, the error happens intermittently with none code adjustments. For this testing surroundings, we observe this firebase information.

The newest subject happens throughout code verification, which returns the error proven beneath. The verification ID is nil, although we’re setting it to an empty string.

The telephone auth credential was created with an empty verification ID.

Under are the primary features we use to register/sign up a person with a telephone quantity.

/// Referred to as the perform to despatched verification code
func sentVerificationCode(to telephone: String) async throws {
    // That is our testing surroundings 
    #if DEV
        Auth.auth().settings?.isAppVerificationDisabledForTesting = true
    #endif
        
    do {
       let verificationID = strive await PhoneAuthProvider.supplier().verifyPhoneNumber(phoneNumber, uiDelegate: nil)
       UserDefaultsStorage.shared.set(key: .phoneVerificationId, string: verificationID)
       return
    } catch {
       throw error
    }
}

/// Referred to as to create the person with the verification code that was obtained
func createUser(with verificationCode: String, telephone: String) async throws {
    // 1 - Get VerificationID
    var verificationID: String? = UserDefaultsStorage.shared.getString(for: .phoneVerificationId)
    #if DEV
    // For DEV we're eradicating the verification utilizing `Auth.auth().settings?.isAppVerificationDisabledForTesting = true`. Which means that the verification code can be nil, so we have to set it to an empty string.
    verificationID = verificationID ?? ""
    #endif

    // 2 - Create credentials
    let credentials = PhoneAuthProvider.supplier().credential(withVerificationID: verificationID, verificationCode: verificationCode)
    // 3 - Sign up
    do {
       let consequence = strive await Auth.auth().signIn(with: credentials)
    } catch let error {
       // Is right here the place Im receiving the error talked about above.
       throw error
    }
}

I’ve examined in a simulator and actual system and the identical consequence.

Testing info:

  • Simulator: iPhone 16 – OS 18.1
  • Actual system: iPhone 13 – OS 18.3.2
  • Xcode 16.3

Andy Nightingale, VP of Product Advertising at Arteris – Interview Collection

0


Andy Nightingale, VP of Product Advertising at Arteris is a seasoned world enterprise chief with a various background in engineering and product advertising and marketing. He’s a Chartered Member of the British Laptop Society and the Chartered Institute of Advertising, and has over 35 years of expertise within the high-tech {industry}.

All through his profession, Andy has held a spread of roles, together with engineering and product administration positions at Arm, the place he spent 23 years. In his present position as VP of product advertising and marketing at Arteris, Andy oversees the Magillem system-on-chip deployment tooling and FlexNoC and Ncore network-on-chip merchandise.

Arteris is a catalyst for system-on-chip (SoC) innovation because the main supplier of semiconductor system IP for the acceleration of SoC growth. Arteris Community-on-Chip (NoC) interconnect mental property (IP) and SoC integration expertise allow larger product efficiency with decrease energy consumption and sooner time to market, delivering confirmed flexibility and higher economics for system and semiconductor corporations, so progressive manufacturers are free to dream up what comes subsequent.

Together with your intensive expertise at Arm and now main product administration at Arteris, how has your perspective on the evolution of semiconductor IP and interconnect applied sciences modified through the years? What key tendencies excite you essentially the most at the moment?

It’s been a unprecedented journey—from my early days writing take a look at benches for ASICs at Arm to serving to form product technique at Arteris, the place we’re on the forefront of interconnect IP innovation. Again in 1999, system complexity quickly accelerated, however the focus was nonetheless totally on processor efficiency and important SoC integration. Verification methodologies had been evolving, however interconnect was typically seen as a hard and fast infrastructure—crucial however not strategic.

Quick-forward to at the moment and interconnect IP has turn into a vital enabler of SoC (System-on-Chip) scalability, energy effectivity, and AI/ML efficiency. The rise of chiplets, domain-specific accelerators, and multi-die architectures has positioned immense stress on interconnect applied sciences to turn into extra adaptive, progressive, bodily, and software-aware.

One of the crucial thrilling tendencies I see is the convergence of AI and interconnect design. At Arteris, we’re exploring how machine studying can optimize NoC (Community-on-Chip) topologies, intelligently route knowledge visitors, and even anticipate congestion to enhance real-time efficiency. This isn’t nearly velocity—it is about making methods extra progressive and responsive.

What excites me is how semiconductor IP is changing into extra accessible to AI innovators. With high-level SoC configuration IP and abstraction layers, startups in automotive, robotics, and edge AI can now leverage superior interconnect architectures while not having a deep background in RTL design. That democratization of functionality is gigantic.

One other key shift is the position of digital prototyping and system-level modeling. Having labored on ESL (Digital System Stage) instruments early in my profession, it’s rewarding to see these methodologies now enabling early AI workload analysis, efficiency prediction, and architectural trade-offs lengthy earlier than silicon is taped out.

In the end, the way forward for AI is determined by how effectively we transfer knowledge—not simply how briskly we course of it. That’s why I imagine the evolution of interconnect IP is central to the following technology of clever methods.

Arteris’ FlexGen leverages AI pushed automation and machine studying to automate NoC (Community-on-Chip) topology technology. How do you see AI’s position evolving in chip design over the following 5 years?

AI is essentially reworking chip design, and over the following 5 years, its position will solely deepen—from productiveness help to clever design accomplice. At Arteris, we’re already dwelling that future with FlexGen, the place AI, formal strategies, and machine studying are central to automating Community-on-Chip (NoC) topology optimization and SoC integration workflows.

What units FlexGen aside is its mix of ML algorithms—all mixed to initialize floorplans from photographs, generate topologies, configure clocks, scale back Clock Area Crossings, and optimize the connectivity topology and its placement and routing bandwidth, streamlining communication between IP blocks. Furthermore, that is all achieved deterministically, which means that outcomes might be replicated and incremental changes made, enabling predictable best-in-class outcomes to be used instances starting from AI help for an skilled SoC designer to creating the appropriate NoC for a novice.

Over the following 5 years, AI’s position in chip design will shift from helping human designers to co-designing and co-optimizing with them—studying from each iteration, navigating design complexity in real-time, and in the end accelerating the supply of AI-ready chips. We see AI not simply making chips sooner however making sooner chips smarter.

The semiconductor {industry} is witnessing speedy innovation with AI, HPC, and multi-die architectures. What are the largest challenges that NoC design wants to resolve to maintain up with these developments?

As AI, HPC, and multi-die architectures drive unprecedented complexity, the largest problem for NoC design is scalability with out sacrificing energy, efficiency, or time to market. At the moment’s chips function tens to a whole bunch of IP blocks, every with totally different bandwidth, latency, and energy wants. Managing this variety—throughout a number of dies, voltage domains, and clock domains—requires NoC options that go far past guide strategies.

NoC resolution applied sciences similar to FlexGen assist handle key bottlenecks: minimizing wire size, maximizing bandwidth, aligning with bodily constraints, and doing every little thing with velocity and repeatability.

The way forward for NoC should even be automation-first and AI-enabled, with instruments that may adapt to evolving floorplans, chipset-based architectures, and late-stage modifications with out requiring full rework. That is the one approach to preserve tempo with trendy SoCs’ large design cycles and heterogeneous calls for and guarantee environment friendly, scalable connectivity on the coronary heart of next-gen semiconductors.

The AI chipset market is projected to develop considerably. How does Arteris place itself to help the growing calls for of AI workloads, and what distinctive benefits does FlexGen supply on this house?

Arteris is just not solely uniquely positioned to help the AI chiplet market however has been doing this already for years by delivering automated, scalable Community-on-Chip (NoC) IP options purpose-built for the calls for of AI workloads together with Generative AI and Massive Language Fashions (LLM) compute —supporting excessive bandwidth, low latency, and energy effectivity throughout more and more advanced architectures.  FlexGen, as the most recent addition to the Arteris NoC IP lineup, will play an much more vital position in quickly creating optimum topologies finest suited to totally different large-scale, heterogeneous SoCs.

FlexGen gives incremental design, partial completion mode, and superior pathfinding to dynamically optimize NoC configurations with out full redesigns—vital for AI chips that evolve all through growth.

Our prospects are already constructing Arteris expertise into multi-die and chiplet-based methods, effectively routing visitors whereas respecting floorplan and clock area constraints on every chiplet. Non-coherent multi-die connectivity is supported over industry-standard interfaces supplied by third- social gathering controllers.

As AI chip complexity grows, so does the necessity for automation, adaptability, and velocity. FlexGen delivers all three, serving to groups construct smarter interconnects—sooner—to allow them to deal with what issues: advancing AI efficiency at scale.

With the rise of RISC-V and customized silicon for AI, how does Arteris’ strategy to NoC design differ from conventional interconnect architectures?

Conventional interconnect architectures had been primarily constructed for fixed-function designs, however at the moment’s RISC-V and customized AI silicon demand a extra configurable, scalable, and automatic strategy than a modified one-size-fits-all resolution. That’s the place Arteris stands aside. Our NoC IP, particularly with FlexGen, is designed to adapt to the variety and modularity of recent SoCs, together with customized cores, accelerators, and chiplets, as talked about above.

FlexGen allows designers to generate and optimize topologies that mirror distinctive workload traits, whether or not low-latency paths for AI inference or high-bandwidth routes for shared reminiscence throughout RISC-V clusters. Not like static interconnects, FlexGen’s algorithms tailor every NoC to the chip’s structure throughout clock domains, voltage islands, and floorplan constraints.

Consequently, Arteris allows groups constructing customized silicon to maneuver sooner, scale back threat, and get essentially the most from their extremely differentiated designs—one thing conventional interconnects weren’t constructed to deal with.

FlexGen claims a 10x enchancment in design iteration velocity. Are you able to stroll us by how this automation reduces complexity and accelerates time-to-market for System-on-Chip (SoC) designers?

FlexGen delivers a 10x enchancment in design iteration velocity by automating among the most advanced and time-consuming duties in NoC design. As an alternative of manually configuring topologies, resolving clock domains, or optimizing routes, designers use FlexGen’s bodily conscious, AI-powered engine to deal with these in hours (or much less)—duties that historically took weeks.

As talked about above, partial completion mode can routinely end even partially accomplished designs, preserving guide intent whereas accelerating timing closure.

The result’s a sooner, extra correct, and easier-to-iterate design movement, enabling SoC groups to discover extra architectural choices, reply to late-stage modifications, and get to market sooner—with higher-quality outcomes and fewer threat of pricey rework.

Certainly one of FlexGen’s standout options is wire size discount, which improves energy effectivity. How does this influence total chip efficiency, notably in power-sensitive functions like edge AI and cell computing?

Wire size straight impacts energy consumption, latency, and total chip effectivity—each in cloud AI / HPC functions that use the extra superior nodes and edge AI inference functions the place each milliwatt issues. FlexGen’s skill to routinely decrease wire size—typically as much as 30%—means shorter knowledge paths, diminished capacitance, and fewer dynamic energy draw.

In real-world phrases, this interprets to decrease warmth technology, longer battery life, and higher performance-per-watt, all of that are vital for AI workloads on the edge or in cell environments and the cloud by straight impacting the overall price of possession (TCO). By optimizing the NoC topology with AI-guided placement and routing, FlexGen ensures that efficiency targets are met with out sacrificing energy effectivity—making it a perfect match for at the moment and tomorrow’s energy-sensitive designs.

Arteris has partnered with main semiconductor corporations in AI knowledge facilities, automotive, shopper, communications, and industrial electronics. Are you able to share insights on how FlexGen is being adopted throughout these industries?

Arteris NoC IP sees sturdy adoption throughout all markets, notably for high-end, extra superior chiplets and SoCs. That’s as a result of it addresses every sector’s prime challenges: efficiency, energy effectivity, and design complexity whereas preserving the core performance and space constraints.

In automotive, for instance, corporations like Dream Chip use FlexGen to hurry up the intersection of AI and Security for autonomous driving by leveraging Arteris for his or her ADAS SoC design whereas assembly strict energy and security constraints. FlexGen’s good NoC optimization and technology in knowledge facilities assist handle large bandwidth calls for and scalability, particularly for AI coaching and total acceleration workloads.

FlexGen offers a quick, repeatable path to optimized NoC architectures for industrial electronics, the place design cycles are tight and product longevity is essential. Clients worth its incremental design movement, AI-based optimization, and talent to adapt shortly to evolving necessities, making FlexGen a cornerstone for next-generation SoC growth.

The semiconductor provide chain has confronted vital disruptions lately. How is Arteris adapting its technique to make sure Community-on-Chip (NoC) options stay accessible and scalable regardless of these challenges?

Arteris responds to produce chain disruptions by doubling down on what makes our NoC options resilient and scalable: automation, flexibility, and ecosystem compatibility.

FlexGen helps prospects design sooner and stay extra agile to regulate to altering silicon availability, node shifts, or packaging methods. Whether or not they’re doing spinoff designs or creating new interconnects from scratch.

We additionally help prospects with totally different course of nodes, IP distributors, and design environments, making certain prospects can deploy Arteris options no matter their foundry, EDA instruments, or SoC structure.

By decreasing dependency on anyone a part of the availability chain and enabling sooner, iterative design, we’re serving to prospects derisk their designs and keep on schedule —even in unsure occasions.

Wanting forward, what are the largest shifts you anticipate in SoC growth, and the way is Arteris getting ready for them?

One of the crucial vital shifts in SoC growth is the transfer towards heterogeneous architectures, chiplet-based designs, and AI-centric workloads. These tendencies demand way more versatile, scalable, and clever interconnects—one thing conventional strategies can’t sustain with.

Arteris is getting ready by investing in AI-driven automation, as seen in FlexGen, and increasing help for multi-die methods, advanced clock/energy domains, and late-stage floorplan modifications. We’re additionally centered on enabling incremental design, sooner iteration, and seamless IP integration—so our prospects can preserve tempo with shrinking growth cycles and rising complexity.

Our objective is to make sure SoC (and chiplet) groups keep agile, whether or not they’re constructing for edge AI, cloud AI, or something in between, all whereas offering one of the best energy, efficiency, and space (PPA) regardless of the complexity of the design, XPU structure, and foundry node used.

Thanks for the nice interview, readers who want to study extra ought to go to Arteris

Create customized Auth on Firebase for SIWA (Sign up with Apple) customers after switch of iOS app to totally different App Retailer Join crew


Apparently, I must create a customized Auth for Firebase — on account of not with the ability to straight modify providerData (also called provider_data) that is a part of the Person Document. Any strategies or steering is appreciated.

Our iOS app makes use of Sign up with Apple, and that labored positive with Firebase’s built-in performance… till we wanted to switch the app to a unique crew in App Retailer Join. It seems that the ID and personal relay e mail deal with that Apple offers for every consumer is exclusive to the App Retailer Join crew, not merely distinctive to the app. The result’s that when attempting to log in with SIWA beneath the brand new crew, Firebase’s regular auth is unable to find the consumer’s account — as a result of the ID from Apple is totally different beneath the brand new crew, and the e-mail deal with will even be totally different if the consumer chosen the personal relay e mail deal with choice.

Firebase shops the sub (the ID offered by Apple) because the uid which is a part of the providerData on the Person Document, however my understanding is that we can’t modify the providerData straight, and so we’ll must create a very customized Auth setup on the Firebase aspect as a way to enable SIWA customers to proceed to log in after our app is transferred to the brand new App Retailer Join crew.

Has anybody gone by way of this course of earlier than? If that’s the case, any suggestions or strategies for how one can setup the customized Auth in Firebase?

I perceive the method from the Apple aspect fairly properly now… as we’re making ready to switch the app from Workforce A to Workforce B, for every consumer we’ll name Apple’s endpoint /auth/usermigrationinfo as Workforce A to retrieve the transfer_sub (also called the “switch identification”) which we then will to retailer someplace in Firebase. Then as Workforce B, we’ll name Apple’s endpoint /auth/usermigrationinfo as a way to retrieve the consumer’s new sub and their probably new e mail deal with… and we’ll once more need to retailer that in Firebase.

The tough half appears to be how one can create the customized Auth performance for Firebase.