Home Blog

From Rejections to Offered-Out Session: My Journey to the Cisco Dwell Stage


For years, I watched Cisco Dwell periods with admiration — every one sparking the identical thought in my thoughts: “In the future, I wish to be up there, too.” I utilized to talk at Cisco Dwell yr after yr, refining concepts, evolving with expertise tendencies, and studying from every “thanks, however not this time” electronic mail, every one a small sting of disappointment, but in addition motivation to maintain pushing ahead. A man holding his Cisco Live badge smiles.A man holding his Cisco Live badge smiles.

In 2025, that dream lastly turned a actuality.

My session, targeted on chatbots and digital assistants in Cisco Contact Centres, was accepted for Cisco Dwell EMEA 2025 — and to my absolute astonishment, it bought out, with a ready record of attendees wanting to dive into the subject.

The Excellent Timing

Cisco Dwell isn’t simply an occasion — it’s a world stage the place expertise leaders and innovators share concepts that form the long run. For me, being accepted as a speaker wasn’t simply an achievement, it was the belief of a dream years within the making. To lastly stand on that stage felt like stepping right into a second I had envisioned for therefore lengthy.

The timing couldn’t have been extra good. Cisco had simply launched the Webex AI Agent, a complicated digital assistant designed to revolutionize buyer assist, and I had the unbelievable alternative to faucet into the joy over this launch with my speak, giving real-world context to a cutting-edge answer.

I shared how AI-powered digital assistants are reworking the best way we take into consideration buyer experiences — automating repetitive duties, enhancing decision instances, and letting people deal with complicated, high-value interactions. However what really amazed me was how the viewers related with the subject. Individuals didn’t simply fill the seats — they had been engaged, sparking conversations and asking questions that prolonged past the session itself.

Extra Than Only a Speak Man standing n front of audience giving a class.Man standing n front of audience giving a class.

The expertise didn’t finish when the session did. Being accepted as a Cisco Dwell speaker — after so many tries — has not solely been a profession milestone however a profession catalyst. It’s positioned me unexpectedly as a Topic Matter Knowledgeable in Contact Centre AI, opening doorways to conversations, collaborations, and new alternatives that weren’t even on my radar earlier than.

However most significantly, it jogged my memory that persistence pays off, particularly while you’re supported by an organization like Cisco.

Why Cisco Makes This Potential

Cisco isn’t only a tech firm — it’s a spot that believes within the individuals behind the concepts. I’m extremely grateful to work in an atmosphere that empowers staff to develop, converse up, and take the stage (actually). Being given the platform to share one thing I’m captivated with — and seeing that keenness resonate with others — is a second I’ll carry ahead with pleasure.

What’s distinctive about Cisco is the way it encourages us to dream large and gives the instruments to make these desires a actuality. To anybody with large profession desires: hold going. Your voice issues, your concepts matter, and at Cisco, they’ll all the time have a spot to shine.

Mine did — and it’s just the start.

Are you able to discover a place the place you’ll be able to shine? Uncover a profession alternative now!

Subscribe to the WeAreCisco Weblog.

Share:

Moore’s Regulation, previous and future – Comfortable Machines


Moore’s Regulation – and the expertise it describes, the built-in circuit – has been one of many defining options of the previous half century. The thought of Moore’s legislation has been invoked in three associated senses. In its unique kind, it was moderately a exact prediction concerning the charge of enhance of the variety of transistors to be fitted on a single built-in circuit. It’s by no means been a legislation – it’s been extra of an organising precept for an trade and its provide chain – and thus a self-fulfilling prophecy. On this sense, it’s been roughly true for 50 years – however is now bumping up in opposition to bodily limits.

Within the second sense, Moore’s legislation is used extra loosely as an announcement concerning the enhance in computing energy, and the discount of its value, over time. The assertion is that computing energy grows exponentially. This additionally was true, for some time. From the mid 1980’s to the mid 2000’s, laptop energy grew at a charge of fifty% a 12 months compounded, doubling each two years. On this extraordinary interval, there was greater than a thousandfold cumulative enhance over a few many years.

The speed of enhance in uncooked laptop energy has slowed considerably during the last 20 years, following the tip of Dennard scaling and the constraints of warmth dissipation, however this has been counteracted to some extent by software program enhancements and the event of architectures specialised for explicit functions. For instance, the Graphics Processing Items – GPUs – which have emerged as being so essential for AI are extremely optimised for multiplying giant matrices.

Within the third sense, Moore’s Regulation is used as a synecdoche for the extra basic thought of accelerating change, that the tempo of change in expertise generally is exponential – and even super-exponential – in character. This in fact is a commonplace in airport enterprise books. It underpins the concept of a forthcoming singularity, as a obtained knowledge in Silicon Valley. The thought of the singularity has been given extra salience by the current fast progress in synthetic intelligence, and the widespread view that superhuman synthetic basic intelligence will quickly be upon us.

On this submit, I need to return to the basics – how a lot the fundamental parts of computing might be shrunk in measurement, and what the prospects for future miniaturisation are. However this does instantly bear on the query of the prospects enhance in laptop energy, which has taken on new significance, as the fabric foundation of the AI increase. AI has introduced us to a brand new scenario; within the classical interval of quickest development of laptop energy (the 80s and 90s) the availability of computing energy was rising exponentially, and the chance was to search out methods of utilizing that energy. Now, with AI, it’s the demand for computing energy that’s rising exponentially, and the problem is whether or not provide can match that demand.

Moore’s Regulation, previous and future – Comfortable Machines

Moore’s Regulation. From Max Roser, Hannah Ritchie, and Edouard Mathieu (2023) – “What’s Moore’s Regulation?” Printed on-line at OurWorldinData.org. Retrieved from: ‘https://ourworldindata.org/moores-law‘ [Online Resource]. Licensed underneath CC-BY.

A classical depiction of Moore’s legislation is proven on this plot from Our World in Information – with a logarithmic y-axis, a straight line signifies an exponential development within the variety of transistors in successive generations of microprocessor. The seemingly inexorable upward progress of the road conceals an enormous quantity of innovation; every upward step was facilitated by analysis and growth of latest supplies and new processes. It additionally conceals some important discontinuities.

For instance, the sooner relationship between laptop energy and variety of transistors was damaged within the mid-2000s. Earlier than then miniaturisation introduced a double profit – it gave you extra transistors on every chip, and as well as every transistor labored quicker, as a result of it was smaller. The latter relation – Dennard scaling – broke down, as a result of warmth dissipation turned a limiting issue

One other elementary change occurred in 2012. The basic unit of the fashionable built-in circuit is the metallic oxide silicon area impact transistor – the mosFET. This consists channel of doped silicon, with contacts at both finish. The channel is coated with a skinny, insulating layer of oxide, on high of which is a metallic electrode – the gate. It’s the gate which controls the move {of electrical} present by the channel. When bodily limits meant that the planar mosFET couldn’t be shrunk any extra, a brand new design flipped the channel into the vertical airplane, so the transistors took the type of fins standing up from the airplane of the silicon chip. Either side of the doped silicon fin is coated by insulating oxide and a metallic gate, to kind the finFET.

The patterns that make the circuits in built-in circuits are made by lithography – gentle is shone by a patterned masks onto a photoresist, which is subsequently developed to make the sample bodily. The decrease restrict on the scale of the options that may be patterned on this manner is in the end set by the wavelength of sunshine used. By way of the 2010’s, lithography was based mostly on utilizing deep ultraviolet gentle created by excimer lasers – with a 193 nm wavelength. By 2020, this method had been squeezed so far as it could go, and the 5 nm course of node makes use of excessive UV, with a wavelength of 13.5 nm. The Dutch firm ASML has a monopoly on the instruments to supply EUV for lithography, every of which prices greater than $100 million; the radiation is created in a metallic plasma, and must be centered totally by mirrors.

I’ve referred to the 2020 iteration of fabrication expertise because the “5 nm course of”, following a long-standing trade conference of characterising successive expertise generations by a single size. Within the days of the planar mosFET, a single parameter characterised the scale of every transistor – the gate size. There was a steady relationship between the gate size and the size characterising the node quantity, and there was a roughly biennial lower within the node quantity, from the 1982 1.5µm course of that drove the explosion of non-public computer systems, to the 2002 90 nm strategy of the Pentium 4. However with the substitute of the mosFET by the finFET, circuit geometry modified and the connection between the node measurement and precise dimensions of the circuit broke down. In actual fact, the node measurement now’s finest regarded as totally a advertising machine, on the precept that the smaller the quantity the higher.

A higher strategy to describe progress within the cutting down of the scale makes use of an estimate of the minimal potential space for a transistor because the product of the metallic pitch, the minimal distance between horizontal interconnects, and the contacted gate pitch, the gap from one transistor’s gate to a different’s.

Minimal transistor footprint (product of metallic pitch and contacted gate pitch) for successive semiconductor course of nodes. Information: (1994 – 2014 inclusive) – Stanford Nanoelectronics Lab, submit 2017 and projections, successive editions of the IEEE Worldwide Roadmap for Gadgets and Programs

My plot exhibits the minimal transistor footprint, calculated on this manner, for every course of node since 1994 (the 350 nm node). The primary 5 nodes – till 2002 – observe the exponential enhance in density anticipated from Moore’s legislation – the match represents a transistor density that doubles each 2.2 years. The final three generations of planar mosFET expertise – till 2009 – present a slight easing of the tempo. The change to the finFET extended the pattern for one more decade or so. But it surely’s clear now that the “2 nm” node, being launched by TSMC this 12 months, confirms a marked levelling off of the tempo of miniaturisation. For this node, there was one other change of geometry – finFETs have been changed by vertical rows of nanowires, every fully surrounded by the metallic of the gate electrode – GAA, for “gate throughout”.

It must be burdened that miniaturisation of transistors is much from the one manner through which laptop energy might be elevated. A superb illustration of this comes from progress in making the ultra-powerful chips which have pushed the present AI increase, comparable to Nvidia’s H100. The H100 itself was truly fabricated by TSMC on the “5 nm” node, the primary to make use of AMSL’s EUV gentle supply for lithography. However, as this text explains, solely a fraction of the efficiency enhancements of the H100 over earlier generations are attributed to Moore’s legislation. A lot of the advance comes from extra environment friendly methods of representing numbers and finishing up the arithmetic operations that underlie synthetic intelligence.

One other issue of rising significance is in the best way particular person silicon chips are packaged. Many fashionable built-in circuits, together with the H100, will not be a single chip. As a substitute a number of particular person chips, together with each logic and reminiscence, are mounted collectively on a silicon substrate, with quick interconnects to affix all of them up. The H100 depends on an TSMC superior packaging expertise generally known as “Chip on Wafer on Substrate” (CoWoS), and is an instance of a “System in Package deal”.

What does the long run maintain? The newest (2023) iteration of the IEEE’s Worldwide Roadmap for Gadgets and Programs foresees another iteration of the Gate All Round structure. The 2031 node is a refinement of that which stacks two mosFETs on high of one another, one with a p-doped channel, one with an n-doped channel (this mixture of p- and n- doped FETs is the elemental unit of logic gates in CMOS expertise – “complementary metallic oxide silicon”, therefore that is known as CFET). This primarily doubles the transistor density. After this, no additional shrinking in dimensions is envisaged, so additional will increase in transistor density are to be obtained by stacking a number of tiers of circuits vertically on the wafer.

So what’s the standing of Moore’s legislation now? I return to the three senses through which individuals discuss Moore’s legislation – as a technical prediction concerning the development within the variety of transistors on an built-in circuit, as a extra basic assertion about growing laptop energy, and as a shorthand for speaking about accelerating technical change generally.

Within the first, and strictest, sense, we might be definitive – Moore’s legislation has run its course. The speed of enhance in transistor density has considerably slowed since 2020, and exponential development with an growing time fixed isn’t exponential any extra. The expertise in its present kind has now begun to hit limits, each bodily and financial.

For the second, looser, sense, issues are extra debatable. Obtainable computing energy remains to be growing, and we see the outcomes of that in advances comparable to the event of huge language fashions. However this elevated energy is coming, much less from miniaturisation, extra from software program, specialised architectures optimised for explicit duties, and superior packaging of chips in “Programs in Package deal”. It’s this transition that underlies the truth that Nvidia is price extra as an organization than TSMC, although it’s TSMC that really manufactures (and packages) the chips.

However I wonder if these approaches shall be topic to diminishing returns, in distinction with the classical interval of Moore’s legislation, when fixed, giant, fractional returns had been repeated 12 months after 12 months for many years, producing orders of magnitude cumulative enhancements. We’re additionally seeing as a serious supply of accelerating laptop energy the brute-force method of simply shopping for increasingly more chips, in large, vitality consuming knowledge centres. These sort of will increase in laptop energy are essentially linear, moderately than exponential, in character, and but they’re making an attempt to satisfy a requirement – largely from AI – which is rising exponentially.

It’s very tempting to take Moore’s legislation as an emblem of the concept technological change generally is accelerating exponentially, however I feel that is unhelpful. Know-how isn’t a single factor that improves at a given charge; there are various applied sciences, and at a given time some shall be accelerating, some shall be stagnating, some could even be regressing. As now we have seen earlier than, the exponential enchancment of a single expertise by no means continues ceaselessly; bodily or financial limits present up, and development saturates. Steady progress wants the continual introduction of latest applied sciences which may take up the baton of development from these older applied sciences, whose development is stalling.

It needs to be burdened right here that once we discuss concerning the finish of Moore’s legislation, the expertise that we’re speaking about isn’t computing generally – it’s this explicit manner of implementing machine logic, CMOS (complementary metallic oxide semiconductor). There are numerous methods through which we are able to think about doing computing – the paradox right here is that CMOS has been so profitable that it has crowded out different approaches, a few of which could have important benefits. For instance, we all know that CMOS logic makes use of a number of orders of magnitude extra vitality per operation than the theoretical minimal (the Landauer restrict).

Lastly, it does bear repeating what a unprecedented interval the heyday of Moore’s legislation and Dennard scaling was, with laptop energy doubling each two years, sustained over a few many years to supply a cumulative thousand-fold enhance. For individuals who have lived by that interval, it will likely be tough to withstand the idea that this charge of technological progress is a part of the pure order of issues.

Snowflake introduces agentic AI improvements for knowledge insights


Snowflake at the moment introduced at its Summit new agentic AI improvements that eradicate the boundaries typically discovered between enterprise knowledge and enterprise exercise, creating AI and ML workflows which are linked and trusted.

Among the many new options is Snowflake Intelligence (public preview quickly), powered by clever knowledge brokers, which supplies a pure language expertise for asking questions that may end up in the supply of actionable insights from structured and unstructured knowledge. Additionally in personal preview quickly is a brand new Information Science Agent to assist knowledge scientists automate routine ML mannequin improvement duties, in response to the corporate’s announcement.

Snowflake Intelligence, powered by LLMs from Anthropic and OpenAI, brings collectively knowledge from quite a few sources and makes use of the brand new  Snowflake Openflow to compile info from spreadsheets, paperwork, photos, and databases concurrently. The information brokers can generate visualizations and help customers in taking motion on insights, Snowflake stated in its announcement. Snowflake Intelligence may also entry third-party information by way of Cortex Data Extensions, quickly to be usually out there on Snowflake Market.

Increasing Snowflake Cortex AI

Snowflake’s AI capabilities are being expanded with the revealing of SnowConvert AI,, an agentic resolution to hurry migrations from legacy platforms to Snowflake, serving to knowledge professionals to modernize their knowledge infrastructures with much less handbook work. A brand new resolution in public preview, Cortex AISQL, makes use of generative AI to assist groups pull insights from a number of knowledge sorts and codecs and construct pipelines utilizing SQL and AI, the corporate stated. This, the corporate wrote in its announcement,  “consists of enriching buyer tables with chat transcripts, correlating sensor readings with inspection photographs, and merging gross sales figures with social media sentiment — enabling analysts to categorise photos, extract insights from name transcripts, and detect anomalies with ease. The result’s a completely built-in SQL expertise throughout all knowledge, unlocking deeper insights, sooner choices, and accelerated innovation with out the necessity for specialised AI abilities or exterior providers. Cortex AISQL delivers unified intelligence throughout your complete group.”

CData Unveils No-Code Snowflake Integration Accelerator

Information connectivity resolution supplier  CData Software program has introduced the launch of the CData Snowflake Integration Accelerator, a knowledge integration software for Snowflake clients. CData stated enterprises can cut back integration construct instances by as much as 90%, enabling sooner insights and maximizing the worth of their Snowflake investments.

The Integration Accelerator supplies no-code knowledge ingestion, transformations and stay connectivity to Snowflake knowledge with three toolkits designed to automation knowledge integration and provides stay entry to Snowflake datas. CData, in its announcement, detailed the brand new instruments as follows:

  • Snowflake Ingestion Toolkit: Speed up knowledge ingestion into Snowflake by 10x with no-code Change Information Seize (CDC) from over 270 sources, together with APIs, databases, and SaaS platforms. This toolkit helps stay integration from gross sales, advertising, monetary, and different enterprise techniques through Snowflake exterior tables, with extra assist for Iceberg Tables and staging.
  • Reside Snowflake Information Entry Toolkit: Allow standards-based stay connectivity to Snowflake from any utility, with native integration capabilities for Microsoft instruments like SSAS, SSRS, and SSIS. This toolkit additionally helps integration with Salesforce by way of Lightning Join, enabling seamless entry to Snowflake knowledge throughout the Salesforce ecosystem.
  • Cortex AI Integration Toolkit: Ingest from structured and semi-structured knowledge with real-time connectivity to energy AI-driven workflows in Snowflake Cortex AI. With instantaneous AI-ready knowledge pipelines from any enterprise supply, Cortex AI has rapid entry to ready knowledge for superior analytics and decision-making. The toolkit additionally helps programmatic orchestration of information pipelines, enabling agentic fashions and integration with Cortex AI’s AI-driven insights, whereas real-time knowledge processing with Change Information Seize (CDC) ensures that Cortex AI operates on probably the most up-to-date knowledge.

 

A Function-Pushed Partnership: How Neighborhood Is At The Heart Of Cisco’s Latest Sports activities Endeavor


After I take into consideration sports activities, I take into consideration group. From the diehard followers {of professional} sports activities groups to leisure gamers, sports activities stay a core a part of society due to how they forge connections and convey individuals collectively. And relating to Cisco’s funding within the sports activities and leisure world, connection is on the coronary heart of every part we do. 

Oftentimes we imply that actually – as Cisco powers a number of the largest occasions in sports activities and leisure, offering safe and dependable know-how to permit followers to observe, share and join in the course of the moments that matter most. From offering on-course Wi-Fi on the U.S. Girls’s Open in Erin Hills this previous weekend, to deploying a full portfolio of options on the U.S. Open at Oakmont in only a few weeks, to elevating the fan expertise at BottleRock Napa Valley, Cisco know-how is protecting the sports activities and leisure world related (and guarded!) week in and week out.  

In the present day, we’re excited to add a brand new associate to our portfolio – Monumental Sports activities & Leisure (MSE). Our new multi-year collaboration will embody the complete Monumental enterprise, together with the NBA’s Washington Wizards, NHL’s Washington Capitals, WNBA’s Washington Mystics, Monumental Sports activities Community (MNMT), and Capital One Enviornment. This new partnership will likely be an unimaginable platform for each our manufacturers – highlighting what turns into doable on the power of Cisco’s AI-ready know-how. However equally as vital, this can be a partnership constructed on shared values – a perception within the energy of sport to function a platform for group empowerment, and the central position that know-how performs in making that perception a actuality.  

The announcement comes at a very thrilling time for MSE, as development is in full swing for the brand-new Capital One Enviornment that’s going to be the brand new gold customary, multi-purpose venue for reside occasions. Within the coronary heart of downtown Washington, D.C, the constructing will serve tens of millions of annual guests and supply an unparalleled platform to focus on our dynamic options, host clients and reveal the optimistic outcomes which can be doable on the again of a unified Cisco infrastructure.  

One of the thrilling options of the brand new Capital One Enviornment is the event of the first-of-its-kind Cisco Command Heart – a state-of-the-art hub throughout the new area that may energy fan experiences, venue operations and workforce efficiency. The Cisco Command Heart will likely be a real showcase for a way an end-to-end Cisco deployment – together with Cisco Wi-Fi 7 with Cisco Areas, Cisco IP Cloth for Media, Cisco ThousandEyes, Wipro VisionEDGE and extra – can energy artistic and dynamic experiences.  

However this partnership is far larger than only a new area. It is a chance to mix the forces of our two organizations and prolong our optimistic affect collectively. One of many methods we’re most excited to leap into that is by means of our work with veterans – a gaggle we’re each extremely keen about serving. This fall, we’re launching a brand new Veterans Management Growth Cohort at Cisco, which is about empowering veterans by offering the instruments, sources, and alternatives they should excel in tech. It is a 10-week profession accelerator program, and we’re excited to host this system fruits with a celebration at Capital One Enviornment this November throughout Veteran’s Week. We hope that that is only a launching pad for one among many community-focused packages that may form this new partnership between Cisco and MSE. I can’t wait to see the brand new heights we are going to attain collectively – showcasing what it means to energy group by means of sports activities.

And if you’re headed to Cisco Stay! subsequent week in San Diego, be sure that to hitch MSE Founder & CEO Ted Leonsis and Brian Tippens SVP, Chief Social Influence and Inclusion Officer at Cisco for a particular session on how this new partnership sits proper on the intersection of Sports activities, Expertise, and Service. 

Share:

Intel spinout Cornelis Networks affords various to Infiniband or Ethernet for HPC and AI networks



The structure incorporates a number of key technical differentiators designed particularly for scale-out parallel computing environments. Credit score-based movement management ensures lossless knowledge transmission, whereas dynamic fine-grained adaptive routing optimizes path choice in real-time. Enhanced congestion management mechanisms are designed to keep up constant efficiency below heavy hundreds, which is a important requirement for AI coaching workloads that may contain hundreds of endpoints.

Efficiency metrics and benchmarking

Cornelis positions the CN5000’s benefits in particular technical metrics that tackle identified bottlenecks in AI and HPC workloads. The corporate claims 2X increased message charges and 35% decrease latency in comparison with different 400Gbps options.

What’s completely different in regards to the Cornelis structure is that with the identical bandwidth, you possibly can obtain double the message charges, Spelman defined. “To me, that’s the way in which that the architectural correctness for the workloads exhibits up.”

For AI workloads particularly, the corporate highlights 6X quicker collective communication efficiency in comparison with distant direct reminiscence entry (RDMA) over converged Ethernet (RoCE) implementations. Collective operations like all-reduce capabilities symbolize important bottlenecks in distributed coaching, the place hundreds of nodes should synchronize gradient updates effectively.

The structure’s congestion administration turns into significantly related in AI coaching eventualities, the place synchronized communication patterns can overwhelm conventional networking approaches. Omni-Path’s credit-based movement management and adaptive routing intention to keep up constant efficiency even below these demanding situations.

“With the very same compute put in and only a swap of the community from one other 400 gig to CN5000, you see utility efficiency that improves by 30%,” Spelman mentioned. “Usually to enhance by 30% on an utility’s efficiency, you would wish a brand new CPU era.”