Settings Results in 4 milliseconds

How to increase Speed of Social Media application
Category: Computer Programming

Speed of the Web Application is really ...


Views: 0 Likes: 40
Android Developer at InnaMed | Y Combinator
Android Developer at InnaMed | Y Combinator

InnaMed is a Y Combinator alum and venture-backed startup developing HomeLab®, a connected, at-home blood testing system for enabling precision medicine. By allowing patients to perform real-time, disease-specific blood tests at home, InnaMed aims to empower healthcare providers and pharmaceutical companies alike to improve and optimize patient therapies. InnaMed’s electrochemical immunoassay and clinical chemistry diagnostics platform is covered by several patents and publications and the company already has multiple ongoing partnerships with top pharmaceutical companies. See our website to learn more (www.InnaMed.com). Our team consists of scientists and engineers from top universities and biotechnology companies. We are passionate about improving upon the cost, access and quality of global healthcare and we are currently looking to hire an Android developer. Specifically, we are looking for an experienced individual to develop Android application(s) for our portable IoT medical diagnostics system. This includes front-end development and integration with the back-end server in a controlled release environment with versioning & source control. You will have independent responsibility and freedom to make your own decisions but will also be collaborating with a cooperative team. Your role will grow with the organization as a core member of the product team and will have a meaningful impact on medicine. Responsibilities 1. Develop Android application architecture for IoT device with features such as 1) integrating hardware sensor readings 2) consuming REST and GraphQL APIs 3) audio-visual-touch UI/UX. 2. Develop and test code to ensure functionality from prototype to commercialization stage 3. Ensure an easy-to-read and documented code base while monitoring and optimizing performance 4. Work with software and hardware engineers and external suppliers to achieve necessary features within an Agile framework Required Skills & Abilities 1. Proficient in front-end Android app development using React Native, Kotlin or other 2. Ability to write clear documentation for QA engineers and non-technical users 3. Creating and expanding code tests 4. Software development in a controlled release environment with versioning & source control 5. Experience working with a modern web stack to develop cloud web applications is a plus 6. Experience working with a quality system in a regulated industry is a plus 7. Attributes enthusiasm, decisiveness, adaptability to changing business requirements, self-starter. 8. Strong written and oral communication skills (English) Required Education and Experience 1. BSc, BEng and/or MSc, PhD in computer science, computer engineering, electrical engineering or related field and a tangible history working in a small team to executed projects from scratch 2. 4+ years of experience in building commercialization ready Android applications


Improve BERT inference speed by combining the power of Optimum, OpenVINO™, ONNX Runtime, and Azure
Improve BERT inference speed by combining the powe ...

In this blog, we will discuss one of the ways to make huge models like BERT smaller and faster with OpenVINO Neural Networks Compression Framework (NNCF) and ONNX Runtime with OpenVINO Execution Provider through Azure Machine Learning.Azure Machine LearningBusiness-critical machine learning models at scale.Learn moreBig models are slow, we need to make them fasterToday’s best-performing language processing models use huge neural architectures with hundreds of millions of parameters. State-of-the-art transformer-based architectures like BERT are available as pretrained models for anyone to use for any language task.The big models have outstanding accuracy, but they are difficult to use in practice. These models are resource hungry due to a large number of parameters. These issues become worse when serving the fine-tuned model and it requires a lot of memory and time to process a single message. A state-of-the-art model is not good if it can handle only one message per second. To improve the throughput, we need to accelerate the well-performing BERT model, by reducing the computation or the number of operations with the help of quantization.Overview of Optimum Intel and quantization aware trainingOptimum Intel is an extension for the Hugging Face Optimum library with OpenVINO runtime as a backend for the Transformers architectures. It also provides an interface to Intel NNCF (Neural Network Compression Framework) package. It helps implement Intel's optimizations through NNCF with changes to just a few lines of code in the training pipeline.Quantization aware training (QAT) is a widely used technique for optimizing models during training. It inserts nodes into the neural network during training that simulates the effect of lower precision. This allows the training algorithm to consider quantization errors as part of the overall training loss that gets minimized during training. QAT has better accuracy and reliability than carrying out quantization after the model has been trained. The output after training with our tool is a quantized PyTorch model, ONNX model, and IR.xml.Overview of ONNXRuntime, and OpenVINO Execution ProviderONNX Runtime is an open source project that is designed to accelerate machine learning across a wide range of frameworks, operating systems, languages, and hardware platforms. It enables the acceleration of machine learning inferencing across all of your deployment targets using a single set of APIs.Intel and Microsoft joined hands to create the OpenVINO Execution Provider (OVEP) for ONNX Runtime, which enables ONNX models for running inference using ONNX Runtime APIs while using the OpenVINO Runtime as a backend. With the OpenVINO Execution Provider, ONNX Runtime delivers better inferencing performance on the same hardware compared to generic acceleration on Intel CPU, GPU, and VPU. Now you've got a basic understanding of quantization, ONNX Runtime, and OVEP, let’s take the best of both worlds and stitch the story together.Putting the tools together to achieve better performanceIn our next steps, we will be doing quantization aware training using Optimum-Intel and Inference using Optimum-ORT with OpenVINO Execution Provider through Azure Machine Learning. Optimum can be used to load optimized models from the Hugging Face Hub and create pipelines to run accelerated inferences.Converting PyTorch FP32 model to INT8 ONNX model with QATWhen utilizing the Hugging Face training pipelines all you need is to update a few lines of code and you can invoke the NNCF optimizations for quantizing the model. The output of this would be an optimized INT8 PyTorch model, ONNX model, and OpenVINO IR. See the flow diagram belowFor this case study, we have chosen the bert-squad pretrained model from Hugging Face. This has been pretrained on the SQuAD dataset for the question-answering use case. QAT can be applied by replacing the Transformers Trainer with the Optimum (OVTrainer). See belowfrom trainer_qa import QuestionAnsweringOVTrainerRun the training pipeline1. Import OVConfigfrom optimum.intel.openvino import OVConfigfrom trainer_qa import QuestionAnsweringOVTrainer2. Initialize a config from the ov_config = OVConfig() 3. Initialize our Trainer trainer = QuestionAnsweringOVTrainer()Comparison of FP32 model and INT8 ONNX model with Netron model visualization toolWhen compared with FP32, the INT8 model has QuantizeLinear and DequantizeLinear operations added to mimic the lower precision after the QAT.Fig1 FP32 modelFig2 INT8 modelTo replicate this example check out the reference code with detailed instructions on QAT and Inference using OpenVINO and Azure Machine Learning Jupyter Notebooks on GitHub.Performance improvement resultsAccuracyOriginal FP32QAT INT8ExplanationF193.192.83In this case, it’s computed over the individual words in the prediction against those in the True Answer. The number of shared words between the prediction and the truth is the basis of the F1 score precision is the ratio of the number of shared words to the total number of words in the prediction, and recall is the ratio of the number of shared words to the total number of words in the ground truth.Eval_exact86.9186.94This metric is as simple as it sounds. For each question + answer pair, if the characters of the model’s prediction exactly match the characters of (one of) the True Answer(s), EM = 1, otherwise EM = 0. This is a strict all-or-nothing metric; being off by a single character results in a score of 0. When assessing against a negative example, if the model predicts any text at all, it automatically receives a 0 for that example.Comparison of ONNXRUNTIME_PERF_TEST application for ONNX-FP32 and ONNX-INT8 modelsWe've used ONNXRuntime APIs for running inference for the BERT model. As you can see the performance for the INT8 optimized model improved almost to 2.95x when compared to FP32 without much compromise in the accuracy.Quantized PyTorch, ONNX, and INT8 models can also be served using OpenVINO Model Server for high-scalability and optimization for Intel solutions so that you can take advantage of all the power of the Intel Xeon processor or Intel's AI accelerators and expose it over a network interface.Optimize speed and performance As neural networks move from servers to the edge, optimizing speed and size becomes even more important. In this blog, we gave an overview of how to use open source tooling to make it easy to improve performance.ReferencesEnhanced Low-Precision Pipeline to Accelerate Inference with OpenVINO toolkit.Developer Guide Model Optimization with the OpenVINO Toolkit. Evaluating QA Metrics, Predictions, and the Null Response.SW/HW configurationFramework configuration ONNXRuntime, Optimum-Intel [NNCF]Application configuration ONNXRuntime, EP OpenVINO ./onnx_perf_test OPENVINO 2022.2 ./benchmark_appInput Question and contextApplication Metric Normalized throughputPlatform Intel Icelake-8380Number of Nodes 2Number of Sockets 2CPU or Accelerator Intel(R) Xeon(R) Platinum 8380 CPU @ 2.30GHzCores/socket, Threads/socket or EU/socket 40,2ucode 0xd000375HT EnabledTurbo EnabledBIOS Version American Megatrends International, LLC. V1.4System DDR Mem Config slots / cap / run-speed 32/32 GB/3200 MT/sTotal Memory/Node (DDR+DCPMM) 1024GBStorage boot INTEL_SSDSC2KB019T8 1.8TNIC 2 x Ethernet Controller X710 for 10GBASE-TOS Ubuntu 20.04.4 LTSKernel 5.15.0-46-genericThe post Improve BERT inference speed by combining the power of Optimum, OpenVINO™, ONNX Runtime, and Azure appeared first on Microsoft Open Source Blog.


How to optimize sql query in Microsoft SQL Server
Category: SQL

1. Keep in mind that when you write a Store Procedure SQL Server generates an SQL plan. If you ha ...


Views: 463 Likes: 102
ASP.NET 8 Best Practices: Coding, Performance Tips ...
Category: .Net 7

In this chapter, we will explore various best practices and performance tips to enhance your ASP. ...


Views: 368 Likes: 98
[Optimize Websites] Add Async to JavaScript to All ...
Category: JavaScript

Optimizing the W ...


Views: 262 Likes: 75
How to Handle Website Scalability
Category: Computer Programming

In layman’s terms, scalability is ...


Views: 0 Likes: 34
Dew Drop – June 21, 2023 (#3969)
Dew Drop – June 21, 2023 (#3969)

Top Links Introducing the New T4 Command-Line Tool for .NET (Mike Corsaro) How to use GitHub Copilot Prompts, tips, and use cases (Rizel Scarlett) How to Hide Your Angular Properties – # vs private Explained (Deborah Kurata) Improved .NET Debugging Experience with Source Link (Patrick Smacchia) 7 Things about C# Running Apps (Joe Mayo) Web & Cloud Development Run OpenTelemetry on Docker (B. Cameron Gain) How Much Will It Hurt? The 10 Things You Need to Do to Migrate Your MVC/Web API App to ASP.NET Core (Peter Vogel) Node v16.20.1 (LTS) and Node v20.3.1 (Current) and Node v18.16.1 (LTS) (Rafael Gonzaga) Service to check if application browser tab is active or not (silfversparre) New W3C website deployed (Coralie Mercier) How to persist Postman variables (Joyce) Dependent Stack Updates with Pulumi Deployments (Komal Ali) Detecting Scene Changes in Audiovisual Content (Avneesh Saluja, Andy Yao & Hossein Taghavi) Exploring the Exciting New Features of TypeScript 5.0 and 5.1 (Suprotim Agarwal) What is an API endpoint? (Postman Team) WinUI, .NET MAUI & XAML .NET MAUI + GitHub Actions + Commas in Certificate Names (Mitchel Sellers) Visual Studio & .NET Integer compression Implementing FastPFor decoding in C# (Oren Eini) Permutations of a String in C# (Matjaz Prtenjak) Using StringBuilder To Replace Values (Khalid Abuhakmeh) Create your own Mediator (like Mediatr) (Steven Giesel) Microsoft Forms Service’s Journey to .NET 6 (Ray Yao) Why is Windows using only even-numbered processors? (Raymond Chen) JetBrains Toolbox App 2.0 Beta Streamlines Installation and Improves Integrations (Victor Kropp) Design, Methodology & Testing One critical skill for a Scrum Master and why? (Martin Hinshelwood) Top 6 AI Coding Assistants in 2023 (Fimber Elemuwa) Big-O Notation and Complexity Analysis (Kirupa Chinnathambi) Cleaning up files changed by a GitHub Action that runs in a container (Rob Bos) To improve as an engineer, get better at requesting (and receiving) feedback (Chelsea Troy) Mobile, IoT & Game Development Get started developing mixed reality for Meta Quest 3 with Unity (Kevin Semple) Screencasts & Videos Technology & Friends – Alex Mattoni on Cycle.io (David Giard) FreeCodeSession – Episode 463 (Jason Bock) What I Wish I Knew… about interviewing for jobs (Leslie Richardson) Podcasts CodeNewbie S24E7 – Navigating Layoffs with Intention (Natalie Davis) (CodeNewbie Team) The Rework Podcast – Buckets of Time (Jason Fried & David Heinemeier Hansson) What It Takes To Be A Web Developer Part 2 – JavaScript Jabber 587 (AJ O’Neal & Dan Shappir) Python Bytes Podcast #341 – Shhh – For Secrets and Shells (Michael Kennedy) Tools and Weapons Podcast – First Vice President Nadia Calviño Architecting Spain’s AI future (Brad Smith) RunAs Radio – Windows Update for Business with Aria Carley (Richard Campbell) Defense Unicorns, A Podcast – Learning from Your Peers with Tracy Gregorio (Rob Slaughter) Community & Events Juneteenth Conference Comes to Chicago (David Giard) Celebrating Tech Trailblazers for Juneteenth (Daniel Ikem) Stack Overflow’s 2023 developer survey Are developers using AI? (Esther Shein) What Does Gen Z Want at Work? The Same Things You Wanted Once Upon a Time (Katie Bartlet) Meet the Skilling Champion Priyesh Wagh (Rie Moriguchi) Things to Do in Philadelphia This Week & Weekend (Visit Philly) The Next Phase of Eleventy Return of the Side Project (Zach Leatherman) Database SQL SERVER – Resolving Deadlock by Accessing Objects in the Same Order (Pinal Dave) The Right Tools for Optimizing Azure SQL Managed Instance Performance (Rie Merritt) Latest features in Azure Managed Instance for Apache Cassandra (Theo van Kraay) T-SQL Tuesday #163 – Career Advice I received (Tracy Boggiano) Miscellaneous Electronic Signatures 2023 Legal Aspects (Bjoern Meyer) Releasing Windows 11 Build 22621.1926 to the Release Preview Channel (Brandon LeBlanc) Windows 11 Moment 3 Heads to the Release Preview Channel (Paul Thurrott) Microsoft CEO Satya Nadella and many Xbox executives are set to defend its FTC case (Tom Warren) More Link Collections The Morning Brew #3731 (Chris Alcock) Sands of MAUI Issue #108 (Sam Basu) Daily Reading List – June 20, 2023 (#107) (Richard Seroter) The Geek Shelf  Learn WinUI 3 (Alvin Ashcraft)


Performance Tuning for ASP.NET Web Applications
Category: .Net 7

Performance Tuning in ASP.NET Core with C# ...


Views: 398 Likes: 109
AI for a cleaner, greener future
AI for a cleaner, greener future

Sure, I'd be happy to help! AI technology has the potential to greatly improve the cleaning and maintenance of public spaces. Here are some examples of how AI can assist in this area1. Autonomous Cleaning Robots One of the most well-known applications of AI in cleaning is autonomous cleaning robots. These robots use sensors, cameras, and machine learning algorithms to navigate and clean public areas such as parks, airports, and shopping centers. For example, the iRobot Roomba i7+ uses AI to map out a room and automatically clean it, while the Dyson Cyclone V10 Absolute uses AI to optimize its cleaning performance based on the surface being cleaned.2. Waste Management AI can also assist in waste management by identifying and sorting different types of waste. For example, the Smartbin system uses AI-powered sensors to detect the type of waste being disposed of and sort it accordingly. This not only makes waste disposal more efficient but also helps reduce contamination and improve recycling rates.3. Predictive Maintenance AI can also be used for predictive maintenance in public spaces. By analyzing data from sensors and other sources, AI algorithms can identify potential issues before they become major problems. For example, the City of Barcelona uses an AI-powered system to monitor its waste management infrastructure and predict when maintenance is needed.4. Energy Management Finally, AI can assist in energy management in public spaces by optimizing lighting and HVAC systems. By analyzing data from sensors and other sources, AI algorithms can identify areas where energy usage is high and adjust the lighting and temperature accordingly. This not only saves energy but also improves comfort and reduces costs for building owners.Overall, AI has the potential to greatly improve the cleaning and maintenance of public spaces. By using autonomous cleaning robots, waste management systems, predictive maintenance, and energy management, cities can create cleaner, greener, and more efficient public areas for everyone to enjoy.


Meet the People Who Use 'Notion' To Plan Their Whole Lives
Meet the People Who Use 'Notion' To Plan Their Who ...

An anonymous reader quotes a report from MIT Technology Review Joshua Bergen is a very productive person. His secret is the workspace app Notion. Bergen, a product manager living in Vancouver, uses it to plan trips abroad in meticulous detail, with notes and timelines. He uses it to curate lists of the movies and TV shows he's watched, and records what he thought of them. It's also a handy way to keep tabs on his 3D-printing projects, map snowboarding runs, and quickly update his cute list of the funny things his kid has said. It might sound strange, but Bergen is one of a growing number of people using Notion, software intended for work, to organize their personal lives. They're using it in a myriad of different ways, from tracking their meditation habits and weekly schedules to logging their water intake and sharing grocery lists. So why has a platform built to accommodate "better, faster work" struck such a chord when there are countless other planning apps out there? Part of the reason Notion has such a devoted fan base is its flexibility. At its heart, Notion is designed to combine the various programs a business might use for functions like HR, sales, and product planning in a single hub. It uses simple templates that let users add or remove features, and remote workers can easily collaborate on notes, databases, calendars, and project boards. This high level of customizability sets Notion apart from other work apps. It's also what's made it so popular among people looking to map out their free time. It started to gain traction around 2018 in YouTube's thriving productivity subculture, where videos of fans swapping time management tips and guides to organizing their lives regularly rack up millions of views. Since then, its following has snowballed. More than 275,000 people have joined a dedicated subreddit, tens of thousands of users share free page templates in private Facebook groups, and TikTok videos advising viewers on how to make their Notion pages look pretty have been watched hundreds of millions of times. "You don't have to change your habits to how rigid software is. The software will change how your mind works," says Akshay Kothari, Notion's cofounder and chief operating officer. "I think that's actually been a big reason why you see so much love in the community because people feel like the things they build are theirs." While platforms like Notion are great for people who enjoy feeling organized, spending too much time optimizing and organizing our lives can be counterproductive when we prioritize creating to-do lists over completing the actual tasks on them, says Gabriele Oettingen, a psychology professor at New York University. It's a phenomenon known as the planning fallacy. Using Notion to track whether you're drinking enough water or going jogging, or using it to plan assignments, doesn't necessarily mean you're actually getting those things done. "In a way, Notion might help me to get structure, but it might not work to get me going," she says. Read more of this story at Slashdot.


An error occurred during the compilation of a reso ...
Category: .Net 7

Question Why is this error happening? "An error occurred during the compilation of a resource re ...


Views: 0 Likes: 33
[Solved] FastCGI, Drupal, Nginx, PHP7.2-FPM Error ...
Category: Servers

When you try to load Drupal Web app on Nginx and Nginx Web server through an error <span style=" ...


Views: 1192 Likes: 97
Full Stack Software Developer
Category: Jobs

We have an opening for a Full Stack Software Developer. Please send resumes asap for our team to ...


Views: 0 Likes: 76
Server-Side Development
Category: Computer Programming

The Se ...


Views: 0 Likes: 20
How do I free up space on linux vm
Category: Research

Title Maximizing Linux Virtual Machine Performance Freeing Up Space and Optimizing Disk Usage< ...


Views: 0 Likes: 0
[IIS] Make IIS Run Faster
Category: Servers

<span style="font-weight bold; text-decoration-line underline; fo ...


Views: 304 Likes: 86
Apply to Android Engineer at Caper
Apply to Android Engineer at Caper

About Us Caper builds smart shopping carts powered by deep learning and computer vision to enable a seamless grab-and-go retail experience. We differ from other emerging cashierless technologies like Amazon Go because we are the scalable solution. Caper’s autonomous checkout technology is plug and play, meaning it requires no in-store renovation, no operational overhaul, no heavy computations or endless image labeling. Any retailer can buy the carts and their entire store is upgraded with cashierless capabilities. Caper costs less than 1% of Amazon Go's infrastructure. We are already live in-stores and our customers love us!  Caper is the fastest-growing company in retail automation technology and is backed by Lux Capital, First Round Capital, Y Combinator along with top executives from Google, Walmart, Instacart, Plated, and Albertsons with over $13M in funding to date. While e-commerce accounts for 8% of total retail spending, Caper is innovating the other 92% of the untapped offline retail potential.  As a part of the Caper team, you’ll be a part of a culture that cares about its people and the future we’re shaping together. At Caper, we may all come from different backgrounds, but we all share one common vision - to fundamentally disrupt the retail industry. We are looking for engineers to join our talented technical team.  Come join our team if...   You MUST have 3+ years of Android development experience Familiarity with common Android / JVM libs like Dagger2, Retrofit, RxJava, and Jetpack Package Common development tools including Git, Linux or Mac cmdline Knowledge of monitoring, optimizing, and profiling of Android applications leveraging tools including LeakCanary, and one or more of the following Lint, Coala, SonarQube, Firebase crashlytics Ability to write clean, modularized code. Understand common Android architecture like MVVM and MVP Vast knowledge of computer science fundamentals such as data structures, operating system, database, networking, as well as domain-driven design experience is preferred Fluency in advanced topics about Adroid, like NDK, Kotlin, etc. Skills to complete unit test and instrumental tests   You are ready to Build and maintain Caper's core application Work closely with the product team to deliver new features and improvements Participate in architecture design Conquer the World! Benefits Competitive salary with equity Excellent medical insurance Unlimited snacks Unlimited PTO


Good Problem Solving Tip
Category: Software Development

<span style="font-size 12pt; font-family Arial; background-color ...


Views: 317 Likes: 108
What is New in Asp Net 7 and EF Core 7 (Best Featu ...
Category: Research

Asp.Net Core 7Navigation If you want the Entity to be included in a ...


Views: 0 Likes: 19
NVIDIA Brings New Generative AI Capabilities, Groundbreaking Performance to 100 Million Windows RTX PCs and Workstations
NVIDIA Brings New Generative AI Capabilities, Grou ...

Generative AI is rapidly ushering in a new era of computing for productivity, content creation, gaming and more. Generative AI models and applications — like NVIDIA NeMo and DLSS 3 Frame Generation, Meta LLaMa, ChatGPT, Adobe Firefly and Stable Diffusion — use neural networks to identify patterns and structures within existing data to generate new and original content. When optimized for GeForce RTX and NVIDIA RTX GPUs, which offer up to 1,400 Tensor TFLOPS for AI inferencing, generative AI models can run up to 5x faster than on competing devices. This is thanks to Tensor Cores — dedicated hardware in RTX GPUs built to accelerate AI calculations — and regular software improvements. Enhancements introduced last week at the Microsoft Build conference doubled performance for generative AI models, such as Stable Diffusion, that take advantage of new DirectML optimizations. As more AI inferencing happens on local devices, PCs will need powerful yet efficient hardware to support these complex tasks. To meet this need, RTX GPUs will add Max-Q low-power inferencing for AI workloads. The GPU will operate at a fraction of the power for lighter inferencing tasks, while scaling up to unmatched levels of performance for heavy generative AI workloads. To create new AI applications, developers can now access a complete RTX-accelerated AI development stack running on Windows 11, making it easier to develop, train and deploy advanced AI models. This starts with development and fine-tuning of models with optimized deep learning frameworks available via Windows Subsystem for Linux. Developers can then move seamlessly to the cloud to train on the same NVIDIA AI stack, which is available from every major cloud service provider. Next, developers can optimize the trained models for fast inferencing with tools like the new Microsoft Olive. And finally, they can deploy their AI-enabled applications and features to an install base of over 100 million RTX PCs and workstations  that have been optimized for AI. “AI will be the single largest driver of innovation for Windows customers in the coming years,” said Pavan Davuluri, corporate vice president of Windows silicon and system integration at Microsoft. “By working in concert with NVIDIA on hardware and software optimizations, we’re equipping developers with a transformative, high-performance, easy-to-deploy experience.” To date, over 400 RTX AI-accelerated apps and games have been released, with more on the way. During his keynote address kicking off COMPUTEX 2023, NVIDIA founder and CEO Jensen Huang introduced a new generative AI to support game development, NVIDIA Avatar Cloud Engine (ACE) for Games. This custom AI model foundry service transforms games by bringing intelligence to non-playable characters through AI-powered natural language interactions. Developers of middleware, tools and games can use ACE for Games to build and deploy customized speech, conversation and animation AI models in their software and games. Generative AI on RTX, Anywhere From servers to the cloud to devices, generative AI running on RTX GPUs is everywhere. NVIDIA’s accelerated AI computing is a low-latency, full-stack endeavor. We’ve been optimizing every part of our hardware and software architecture for many years for AI, including fourth-generation Tensor Cores — dedicated AI hardware on RTX GPUs. Regular driver optimizations ensure peak performance. The most recent NVIDIA driver, combined with Olive-optimized models and updates to DirectML, delivers significant speedups for developers on Windows 11. For example, Stable Diffusion performance is improved by 2x compared to the previous interference times for developers taking advantage of DirectML optimized paths. And with the latest generation of RTX laptops and mobile workstations built on the NVIDIA Ada Lovelace architecture, users can take generative AI anywhere. Our next-gen mobile platform brings new levels of performance and portability — in form factors as small as 14 inches and as lightweight as about three pounds. Makers like Dell, HP, Lenovo and ASUS are pushing the generative AI era forward, backed by RTX GPUs and Tensor Cores. “As AI continues to get deployed across industries at an expected annual growth rate of over 37% now through 2030, businesses and consumers will increasingly need the right technology to develop and implement AI, including generative AI. Lenovo is uniquely positioned to empower generative AI spanning from devices to servers to the cloud, having developed products and solutions for AI workloads for years. Our NVIDIA RTX GPU-powered PCs, such as select Lenovo ThinkPad, ThinkStation, ThinkBook, Yoga, Legion and LOQ devices, are enabling the transformative wave of generative AI for better everyday user experiences in saving time, creating content, getting work done, gaming and more.” — Daryl Cromer, vice president and chief technology officer of PCs and Smart Devices at Lenovo “Generative AI is transformative and a catalyst for future innovation across industries. Together, HP and NVIDIA equip developers with incredible performance, mobility and the reliability needed to run accelerated AI models today, while powering a new era of generative AI.” —  Jim Nottingham, senior vice president and general manager of Z by HP “Our recent work with NVIDIA on Project Helix centers on making it easier for enterprises to build and deploy trustworthy generative AI on premises. Another step in this historic moment is bringing generative AI to PCs. Think of app developers looking to perfect neural network algorithms while keeping training data and IP under local control. This is what our powerful and scalable Precision workstations with NVIDIA RTX GPUs are designed to do. And as the global leader in workstations, Dell is uniquely positioned to help users securely accelerate AI applications from the edge to the datacenter.” — Ed Ward, president of the client product group at Dell Technologies “The generative AI era is upon us, requiring immense processing and fully optimized hardware and software. With the NVIDIA AI platform, including NVIDIA Omniverse, which is now preinstalled on many of our products, we are excited to see the AI revolution continue to take shape on ASUS and ROG laptops.” — Galip Fu, director of global consumer marketing at ASUS Soon, laptops and mobile workstations with RTX GPUs will get the best of both worlds. AI inference-only workloads will be optimized for Tensor Core performance while keeping power consumption of the GPU as low as possible, extending battery life and maintaining a cool, quiet system. The GPU can then dynamically scale up for maximum AI performance when the workload demands it. Developers can also learn how to optimize their applications end-to-end to take full advantage of GPU-acceleration via the NVIDIA AI for accelerating applications developer site.


Login to Continue, We will bring you back to this content 0



For peering opportunity Autonomouse System Number: AS401345 Custom Software Development at ErnesTech Email Address[email protected]