LightBlog

lundi 16 décembre 2019

How Qualcomm Brought Tremendous Improvements in AI Performance to the Snapdragon 865

It seems like we can’t go a day without seeing “artificial intelligence” in the news, and this past week was no exception in no small part thanks to the Snapdragon Tech Summit. Every year, Qualcomm unveils the plethora of improvements it brings to its Hexagon DSP and the Qualcomm AI Engine, a term they use for their entire heterogeneous compute platform – CPU, GPU, and DSP – when talking about AI workloads. A few years ago, Qualcomm’s insistence on moving the conversation away from traditional talking points, such as year-on-year CPU performance improvements, seemed a bit odd. Yet in 2019 and with the Snapdragon 865, we see that heterogeneous computing is indeed at the helm of their mobile computing push, as AI and hardware-accelerated workloads seem to sneak their way into a breadth of use cases and applications, from social media to everyday services.

The Snapdragon 865 is bringing Qualcomm’s 5th generation AI engine, and with it come juicy improvements in performance and power efficiency — but that’s to be expected. In a sea of specifications, performance figures, fancy engineering terms, and tiresome marketing buzzwords, it’s easy to lose sight of what these improvements actually mean. What do they describe? Why are these upgrades so meaningful to those implementing AI in their apps today, and perhaps more importantly, to those looking to do so in the future?

In this article, we’ll take an approachable yet thorough tour of the Qualcomm AI Engine combing through its history, its components and the Snapdragon 865’s upgrades, and most importantly, why or how each of these have contributed to today’s smartphone experience, from funny filters to digital assistants.

The Hexagon DSP and Qualcomm AI Engine: When branding makes a difference

While I wasn’t able to attend this week’s Snapdragon Tech Summit, I have nonetheless attended every other one since 2015. If you recall, that was the year of the hot mess that was the Snapdragon 810, and so journalists at that Chelsea loft in New York City were eager to find out how the Snapdragon 820 would redeem the company. And it was a great chipset, alright: It promised healthy performance improvements (with none of the throttling) by going back to the then-tried-and-true custom cores Qualcomm was known for. Yet I also remember a very subtle announcement that, in retrospect, ought to have received more attention: the second generation Hexagon 680 DSP and its single instruction, multiple data (SIMD) Hexagon Vector eXtensions, or HVX. Perhaps if engineers hadn’t named the feature, it would have received the attention it deserved.

This coprocessor allows the scalar DSP unit’s hardware threads to access HVX “contexts” (register files) for wide vector processing capabilities. It enabled the offloading of significant compute workloads from the power-hungry CPU or GPU to the power-efficient DSP so that imaging and computer vision tasks would run at substantially improved performance per milliwatt. They are perfect for applying identical operations on contiguous vector elements (originally just integers), making them a good fit for computer vision workloads. We’ve written an in-depth article on the DSP and HVX in the past, noting that the HVX architecture lends itself well to parallelization and, obviously, processing large input vectors. At the time, Qualcomm promoted both the DSP and HVX almost exclusively by describing the improvements they would bring to computer vision workloads such as the Harris corner detector and other sliding window methods.

It wasn’t until the advent of deep learning in consumer mobile applications that the DSP, its vector processing units (and now, a tensor accelerator) would get married to AI and neural networks, in particular. But looking back, it makes perfect sense: The digital signal processor (DSP) architecture, originally designed for handling digitized real-world or analog signal inputs, lends itself to many of the same workloads as many machine learning algorithms and neural networks. For example, DSPs are tailored for filter kernels, convolution and correlation operations, 8-bit calculations, a ton of linear algebra (vector and matrix products) and multiply-accumulate (MAC) operations, all most efficient when parallelized. A neural network’s runtime is also highly dependent on multiplying large vectors, matrices and/or tensors, so it’s only natural that the DSP’s performance advantages neatly translate to neural network architectures as well. We will revisit this topic in short!

In subsequent years, Qualcomm continued to emphasize that they offer not just chipsets, but mobile platforms, and that they focus not just on improving particular components, but delivering “heterogeneous” compute. In 2017, they released their Snapdragon Neural Processing Engine SDK (for runtime acceleration) on the Qualcomm Developer Network, and in early 2018 they announced the Qualcomm Artificial Intelligence Engine to consolidate their several AI-capable hardware (CPU, GPU, DSP) and software components under a single name. With this useful nomenclature, they were able to neatly advertise their AI performance improvements on both the Snapdragon 855 and Snapdragon 865, being able to comfortably spell out the number of trillions of operations per second (TOPS) and year-on-year percentage improvements. Harnessing the generational improvements in CPU, GPU, and DSP – all of which see their own AI-focused upgrades – the company is able to post impressive benchmarks against competitors, which we’ll go over shortly. With the company’s recent marketing efforts and unified, consistent messaging on heterogeneous computing, their AI branding is finally gaining traction among journalists and tech enthusiasts.

Demystifying Neural Networks: A mundane pile of linear algebra

To disentangle a lot of jargon we’ll come across later in the article, we need a short primer on what a neural network is and what you need to make it faster. I want to very briefly go over some of the mathematical underpinnings of neural networks, avoiding as much jargon and notation as possible. The purpose of this section is simply to identify what a neural network is doing, fundamentally: the arithmetic operations it executes, rather than the theoretical basis that justifies said operations (that is far more complicated!). Feel free to proceed to the next section if you want to jump straight to the Qualcomm AI Engine upgrades.

“Vector math is the foundation of deep learning.” – Travis Lanier, Senior Director of Product Management at Qualcomm at the 2017 Snapdragon Tech Summit

Below you will find a very typical feedforward fully-connected neural network diagram. In reality, the diagram makes the whole process look a bit more complicated than it is (at least, until you get used to it). We will compute a forward pass, which is ultimately what a network is doing whenever it produces an inference, a term we’ll encounter later in the article as well. At the moment, we will only concern ourselves with the machine and its parts, with brief explanations of each component.

A neural network consists of sequential layers, each comprised of several “neurons” (depicted as circles in the diagram) connected by weights (depicted as lines in the diagram). In general terms, there are three kinds of layers: the input layer, which takes the raw input; hidden layers, which compute mathematical operations from the previous layer, and the output layer, which provides the final predictions. In this case, we have only one hidden layer, with three hidden units. The input consists of a vector, array, or list of numbers of a particular dimension or length. In the example, we will have a two-dimensional input, let’s say [1.0, -1.0]. Here, the output of the network consists of a scalar or single number (not a list). Each hidden unit is associated with a set of weights and a bias term, shown alongside and below each node. To calculate the weighted sum output of a unit, each weight is multiplied with each corresponding input, and then the products are added together. Then, we will simply add the bias term to that sum of products, resulting in the output of the neuron. For example, with our input of [1.0,-1.0], the first hidden unit will have an output of 1.0*0.3 + (-1.0) * 0.2 + 1.0 = 1.1. Simple, right?

The next step in the diagram represents an activation function, and is what will allow us to produce the output vector of each hidden layer. In our case, we will be using the very popular and extremely simple rectified linear unit or ReLU, which will take an input number and output either (i) zero, if that number is negative or zero (ii) the input number itself, if the number is positive. For example, ReLU(-0.1) = 0, but ReLU(0.1) = 0.1. Following the example of our input as it propagates through that first hidden unit, the output of 1.1 that we computed would be passed into the activation function, yielding ReLU(1.1)=1.1. The output layer, in this example, will function just like a hidden unit: it will multiply the hidden units’ outputs against its weights, and then add its bias term of 0.2. The last activation function, the step function, will turn positive inputs into 1 and negative values into 0. Knowing how each of the operations in the network operates, we can write down the complete computation of our inference as follows:

That is all there is to our feedforward neural network computation. As you can see, the operations consist almost entirely of products and sums of numbers. Our activation function ReLU(x) can be implemented very easily as well, for example by simply calling max(x,0), such that it returns x whenever the input is greater than 0, but otherwise it returns 0. Note that step(x) can be computed similarly. Many more complicated activation functions exist, such as the sigmoidal function or the hyperbolic tangent, involving different internal computations and better-suited for different purposes. Another thing you can already begin noticing is that we also can run the three hidden units’ computations, and their ReLU applications, in parallel, as their values are not needed at the same time up until we calculate their weighted sum at the output node.

But we don’t have to stop there. Above, you can see the same computation, but this time represented with matrix and vector multiplication operations instead. To arrive at this representation, we “augment” our input vector by adding a 1.0 to it (lighter hue), such that when we put our weights and our bias (lighter hue) in the matrix as shown above, the resulting multiplication yields the same hidden unit outputs. Then, we can apply ReLU on the output vector, element-wise, and then “augment” the ReLU output to multiply it by the weights and bias of our output layer. This representation greatly simplifies notation, as the parameters (weights and biases) of an entire hidden layer can be tucked under a single variable. But most importantly for us, it makes it clear that the inner computations of the network are essentially matrix and vector multiplication or dot products. Given how the size of these vectors and matrices scale with the dimensionality of our inputs and the number of parameters in our network, most runtime will be spent doing these sorts of calculations. A bunch of linear algebra!

Our toy example is, of course, very limited in scope. In practice, modern deep learning models can have tens if not hundreds of hidden layers, and millions of associated parameters. Instead of our two-dimensional vector input example, they can take in vectors with thousands of entries, in a variety of shapes, such as matrices (like single-channel images) or tensors (three-channel RGB images). There is also nothing stopping our matrix representation from taking in multiple inputs vectors at once, by adding rows to our original input. Neural networks can also be “wired” differently than our feedforward neural network, or execute different activation functions. There is a vast zoo of network architectures and techniques, but in the end, they mostly break down to the same parallel arithmetic operations we find in our toy example, just at a much larger scale.

Visual example of convolution layers operating on a tensor. (Image credit: Towards Data Science)

For example, the popular convolutional neural networks (CNNs) that you likely have read about are not “fully-connected” like our mock network. The “weights” or parameters of its hidden convolutional layers can be thought of as a sort of filter, a sliding window applied sequentially to small patches of an input as shown above — this “convolution” is really just a sliding dot product! This procedure results in what’s often called a feature map. Pooling layers reduce the size of an input or a convolutional layer’s output, by computing the maximum or average value of small patches of the image. The rest of the network usually consists of fully-connected layers, like the ones in our example, and activation functions like ReLU. This is often used for feature extraction in images where early convolutional layers’ feature maps can “detect” patterns such as lines or edges, and later layers can detect more complicated features such as faces or complex shapes.

All of what’s been said is strictly limited to inference, or evaluating a neural network after its parameters have been found through training which is a much more complicated procedure. And again, we’ve excluded a lot of explanations. In reality, each of the network’s components is included for a purpose. For example, those of you who have studied linear algebra can readily observe that without the non-linear activation functions, our network simplifies to a linear model with very limited predictive capacity.

An Upgraded AI Engine on the Snapdragon 865 – A Summary of Improvements

With this handy understanding of the components of a neural network and their mathematical operations, we can begin to understand exactly why hardware acceleration is so important. In the last section, we can observe that parallelization is vital to speeding up the network given it allows us, for example, to compute several parallel dot-products corresponding to each neuron activation. Each of these dot-products is itself constituted of multiply-add operations on numbers, usually with 8-bit precision in the case of mobile applications, that must happen as quickly as possible. The AI Engine offers various components to offload these tasks depending on the performance and power efficiency considerations of the developer.

A diagram of a CNN for the popular MNIST dataset, shown on stage at this year’s Snapdragon Summit. The vector processing unit is a good fit for the fully-connected layers, like in our mock example. Meanwhile, the tensor processor handles the convolutional and pooling layers that process multiple sliding kernels in parallel, like in the diagram above, and each convolutional layer might output many separate feature maps.

First, let’s look at the GPU, which we usually speak about in the context of 3D games. The consumer market for video games has stimulated development in graphics processing hardware for decades, but why are GPUs so important for neural networks? For starters, they chew through massive lists of 3D coordinates of polygon vertices at once to keep track of an in-game world state. The GPU must also perform gigantic matrix multiplication operations to convert (or map) these 3D coordinates onto 2D planar, on-screen coordinates, and also handle the color information of pixels in parallel. To top it all off, they offer high memory bandwidth to handle the massive memory buffers for the texture bitmaps overlaid onto the in-game geometry. Its advantages in parallelization, memory bandwidth, and resulting linear algebra capabilities match the performance requirements of neural networks.

The Adreno GPU line thus has a big role to play in the Qualcomm AI Engine, and on stage, Qualcomm stated that this updated component in the Snapdragon 865 enables twice as much floating-point capabilities and twice the number of TOPS compared to the previous generation, which is surprising given that they only posted a 25% performance uplift for graphics rendering. Still, for this release, the company boasts a 50% increase in the number of arithmetic logic units (ALUs), though as per usual, they have not disclosed their GPU frequencies. Qualcomm also listed mixed-precision instructions, which is just what it sounds like: different numerical precision across operations in a single computational method.

Adreno 650 GPU in the Qualcomm Snapdragon 865

The Hexagon 698 DSP is where we see a huge chunk of the performance gains offered by the Snapdragon 865. This year, the company has not communicated improvements in their DSP’s vector eXtensions (whose performance quadrupled in last year’s 855), nor their scalar units. However, they do note that for this block’s Tensor Accelerator, they’ve achieved four times the TOPs compared to the version introduced last year in the Hexagon 695 DSP, while also being able to offer 35% better power efficiency. This is a big deal considering the prevalence of convolutional neural network architectures in modern AI use cases ranging from image object detection to automatic speech recognition. As explained above, the convolution operation in these networks produces a 2D array of matrix outputs for each filter, meaning that when stacked together, the output of a convolution layer is a 3D array or tensor.

Qualcomm also promoted their “new and unique” deep learning bandwidth compression technique, which can apparently compress data losslessly by around 50%, in turn moving half the data and freeing up bandwidth for other parts of the chipset. It should also save power by reducing that data throughput, though we weren’t given any figures and there ought to be a small power cost to compressing the data as well.

On the subject of bandwidth, the Snapdragon 865 supports LPDDR5 memory, which will also benefit AI performance as it will increase the speed at which resources and input data are transferred. Beyond hardware, Qualcomm’s new AI Model Efficiency Toolkit makes easy model compression and resulting power efficiency savings available to developers. Neural networks often have a large number of “redundant” parameters; for example, they may make hidden layers wider than they need to be. One of the AI Toolkit features discussed on stage is thus model compression, with two of the cited methods being spatial singular value decomposition (SVD) and bayesian compression, both of which effectively prune the neural network by getting rid of redundant nodes and adjusting the model structure as required. The other model compression technique presented on stage relates to quantization, and that involves changing the numerical precision of weight parameters and activation node computations.

The numerical precision of neural network weights refers to whether the numerical values used for computation are stored, transferred, and processed as 64, 32, 16 (half-precision) or 8-bit values. Using lower numerical precision (for example, INT8 versus FP32) reduces overall memory usage and data transfer speeds, allowing for higher bandwidth and faster inferences. A lot of today’s deep learning applications have switched to 8-bit precision models for inference, which might sound surprising: wouldn’t higher numerical accuracy enable more “accurate” predictions in classification or regression tasks? Not necessarily; higher numerical precision, particularly during inference, may be wasted as neural networks are trained to cope with noisy inputs or small disturbances throughout training anyway, and the error on the lower-bit representation of a given (FP) value is uniformly ‘random’ enough. In a sense, the low-precision of the computations is treated by the network as another source of noise, and the predictions remain usable. Heuristic explainers aside, it is likely you will accrue an accuracy penalty when lousily quantizing a model without taking into account some important considerations, which is why a lot of research goes into the subject

Back to the Qualcomm AI Toolkit: Through it they offer data-free quantization, allowing models to be quantized without data or parameter fine-tuning while still achieving near-original model performance on various tasks. Essentially, it adapts weight parameters for quantization and corrects for the bias error introduced when switching to lower precision weights. Given the benefits incurred by quantization, automating the procedure under an API call would simplify model production and deployment, and Qualcomm claims more than four times the performance per watt when running the quantized model.

But again, this isn’t shocking: quantizing models can offer tremendous bandwidth and storage benefits. Converting a model to INT8 not only nets you a 4x reduction in bandwidth, but also the benefit of faster integer computations (depending on the hardware). It is a no-brainer, then, that hardware-accelerated approaches to both the quantization and the numerical computation would yield massive performance gains. On his blog, for example, Google’s Pete Warden wrote that a collaboration between Qualcomm and Tensorflow teams enables 8-bit models to run up to seven times faster on the HVX DSP than on the CPU. It’s hard to overstate the potential of easy-to-use quantization, particularly given how Qualcomm has focused on INT8 performance.

The Snapdragon 865’s ARM-based Kryo CPU is still an important component of the AI engine. Even though the hardware acceleration discussed in the above paragraphs is preferable, sometimes we can’t avoid applications that do not properly take advantage of these blocks, resulting in CPU fallback. In the past, ARM had introduced specific instruction sets aimed at accelerating matrix- and vector-based calculations. In ARMv7 processors, we saw the introduction of ARM NEON, a SIMD architecture extension enabling DSP-like instructions. And with the ARMv8.4-A microarchitecture, we saw the introduction of an instruction specifically for dot-products.

All of these posted performance gains relate to many of the workloads we described in the previous section, but it’s also worth keeping in mind that these Snapdragon 865 upgrades are only the latest improvements in Qualcomm’s AI capabilities. In 2017, we documented their tripling of AI capabilities with the Hexagon 685 DSP and other chipset updates. Last year, they introduced their tensor accelerator, and integrated support for non-linearity functions (like the aforementioned ReLU!) at the hardware level. They also doubled the number of vector accelerators and improved the scalar processing unit’s performance by 20%. Pairing all of this with enhancements on the CPU side, like those faster dot-product operations courtesy of ARM, and the additional ALUs in the GPU, Qualcomm ultimately tripled raw AI capabilities as well.

Practical Gains and Expanded Use-Cases

All of these upgrades have lead to five times the AI capabilities on the Snapdragon 865 compared to just two years ago, but perhaps most importantly, the improvements also came with better performance per milliwatt, a critical metric for mobile devices. At the Snapdragon Summit 2019, Qualcomm gave us a few benchmarks comparing their AI Engine against two competitors on various classification networks. These figures look to be collected using AIMark, a cross-platform benchmarking application, which enables comparisons against Apple’s A-series and Huawei’s HiSilicon processors. Qualcomm claims that these results make use of the entire AI Engine, and we’ll have to wait until more thorough benchmarking to properly disentangle the effect of each component and determine how these tests were conducted. For example, do the results from company B indicate CPU fallback? As far as I’m aware, AIMark currently doesn’t advantage of the Kirin 990’s NPU on our Mate 30 Pro units, for example. But it does support the Snapdragon Neural Processing Engine, so it will certainly take advantage of the Qualcomm AI Engine; given it is internal testing, it’s not explicitly clear whether the benchmark is properly utilizing the right libraries or SDK for its competitors.

It must also be said that Qualcomm is effectively comparing the Snapdragon 865’s AI processing capabilities against previously-announced or released chipsets. It is very likely that its competitors will bring similarly-impactful performance improvements in the next cycle, and if that’s the case, then Qualcomm would only hold the crown for around half a year from the moment Snapdragon 865 devices hit the shelves. That said, these are still indicative of the kind of bumps we can expect from the Snapdragon 865. Qualcomm has generally been very accurate when communicating performance improvements and benchmark results of upcoming releases.

Qualcomm Snapdragon 865 AI performance versus competitors

All of the networks presented in these benchmarks are classifying images from databases like ImageNet, receiving them as inputs and outputting one out of hundreds of categories. Again, they rely on the same kinds of operations we described in the second section, though their architectures are a lot more complicated than these examples and they’ve been regarded as state of the art solutions at their time of publication. In the best of cases, their closest competitor provides less than half the number of inferences per second.

AI power consumption on the Qualcomm Snapdragon 865

In terms of power consumption, Qualcomm offered inferences per watt figures to showcase the amount of AI processing possible in a given amount of power. In the best of cases (MobileNet SSD), the Snapdragon AI Engine can offer double the number of inferences under the same power budget.

Power is particularly important for mobile devices. Think, for example, of a neural network-based Snapchat filter. Realistically, the computer vision pipeline extracting facial information and applying a mask or input transformation only needs to run at a rate of 30 or 60 completions per second to achieve a fluid experience. Increasing raw AI performance would enable you to take higher-resolution inputs and output better looking filters, but it might also simply be preferable to settle for HD resolution for quicker uploads and decrease power consumption and thermal throttling. In many applications, “faster” isn’t necessarily “better”, and one then gets to reap the benefits of improved power efficiency.

Snapdragon acceleration on the Qualcomm Snapdragon 865

During Day 2 of the Snapdragon Summit, Sr. Director of Engineering at Snapchat Yurii Monastyrshyn took the stage to show how their latest deep learning-based filters are greatly accelerated by Hexagon Direct NN using the Hexagon 695 DSP on the Snapdragon 865.

On top of that, as developers get access to easier neural network implementations and more applications begin employing AI techniques, concurrency use cases will take more of a spotlight as the smartphone will have to handle multiple parallel AI pipelines at once (either for a single application processing input signals from various sources or as many applications run separately on-device). While we see respectable power efficiency gains across the compute DSP, GPU, and CPU, the Qualcomm Sensing Hub handles always-on use cases to listen for trigger words at very low power consumption. It enables monitoring audio, video and sensor feeds at under 1mA of current, allowing the device to spot particular sound cues (like a baby crying), on top of the familiar digital assistant keywords. On that note, the Snapdragon 865 enables detecting not just the keyword but also who is speaking it, to identify an authorized user and act accordingly.

More AI on Edge Devices

These improvements can ultimately translate into tangible benefits for your user-experience. Services that involve translation, object recognition and labeling, usage predictions or item recommendations, natural language understanding, speech parsing and so on will gain the benefit of operating faster and consuming less power. Having a higher compute budget also enables the creation of new use cases and experiences, and moving processes that used to take place in the cloud onto your device. While AI as a term has been used in dubious, deceiving and even erroneous ways in the past (even by OEMs), many of your services you enjoy today ultimately rely on machine learning algorithms in some form or another.

But beyond Qualcomm, other chipset makers have been quickly iterating and improving on this front too. For example, the 990 5G brought a 2+1 NPU core design resulting in up to 2.5 times the performance of the Kirin 980, and twice that of the Apple A12. When the processor was announced, it was shown to offer up to twice the frames (inferences) per second of the Snapdragon 855 at INT8 MobileNet, which is hard to square with the results provided by Qualcomm. The Apple A13 Bionic, on the other hand, reportedly offered to six times faster matrix multiplication over its predecessor and improved its eight-core neural engine design. We will have to wait until we can properly test the Snapdragon 865 on commercial devices against its current and future competitors, but it’s clear that competition in this space never stays still as the three companies have been pouring a ton of resources into bettering their AI performance.

The post How Qualcomm Brought Tremendous Improvements in AI Performance to the Snapdragon 865 appeared first on xda-developers.



from xda-developers https://ift.tt/34rDl55
via IFTTT

[Update 6: Los Angeles] Verizon 5G is Rolling Out to More Cities

Update 6 (12/16/19 @ 10:40 AM ET): Verizon is rolling out 5G coverage in the Los Angeles area.

Update 5 (11/20/19 @ 9:10 AM ET): Verizon finally has detailed 5G coverage maps for every city on its website.

Update 4 (11/19/19 @ 9:25 AM ET): Verizon’s 5G network lights up in Boston, Houston, and Sioux Falls.

Update 3 (10/25/19 @ 12:45 PM ET): Verizon expands its 5G network coverage to Omaha and Dallas.

Update 2 (9/26/19 @ 1:15 PM ET): Verizon launches 5G service in New York City, Boise, and Panama City.

Update 1 (8/22/19 @ 12:15 PM ET): Verizon has announced the 5G rollout in Phoenix and a partnership with Boingo.

While many people are still skeptical about 5G, Verizon continues their rollout plans. Today, the company flipped the switch for four new cities: Atlanta, Detroit, Indianapolis, and Washington DC. Verizon is already selling a couple of 5G devices, but the list of available cities is still relatively small. So the continued expansion is good news.

Verizon’s 5G Ultra Wideband network is mmWave, just like AT&T, but different from Sprint’s sub-6Ghz network. One of the limitations of mmWave is you have to be in very specific locations to get the advertised 5G speeds. For example, read the description for Indianapolis below.

Indianapolis:

In Indianapolis, 5G Ultra Wideband service is initially available in parts of the following neighborhoods, Arsenal Heights, Bates Hendricks, Castleton, Crown Hill, Fountain Square, Grace Tuxedo Park, Hawthorne, Historic Meridian Park, Lockerbie Square, Ransom Place, Renaissance Place, St. Joseph Historic Neighborhood, Upper Canal and Woodruff Place and around such landmarks and public spaces as Garfield Park, and Indiana University School of Medicine.

Even if you have a 5G device and live in these cities, you may not be in the covered areas. These four new cities bring Verizon’s list up to nine, but they are still planning to have 5G in more than 30 cities by 2020. Soon, they will add the Galaxy Note 10 5G to the list of capable devices as well. Whether the market is ready or not, Verizon marches on with 5G.

Washington DC:

In Washington DC, consumers, businesses and government agencies can initially access Verizon’s 5G Ultra Wideband service in areas of Foggy Bottom, Dupont Circle, Cardozo / U Street, Adams Morgan, Columbia Heights, Le Droit Park, Georgetown Waterfront, Judiciary Square, Shaw, Eckington, NOMA, National Mall and the Smithsonian, Gallery Place / Chinatown, Mt. Vernon Square, Downtown, Penn Quarter, Brentwood, Southwest Waterfront, Navy Yard, and nearby Crystal City, VA, as well as around landmarks such as the Ronald Reagan National Airport, United States Botanical Gardens, Hart Senate Building, National Gallery of Art, Lafayette Square, The White House, Freedom Plaza, Farragut Square, George Washington University, Capital One Arena, Union Station, Howard University Hospital, George Washington University Hospital, and Georgetown Waterfront Park.

Atlanta:

In Atlanta, 5G Ultra Wideband service will initially be concentrated in parts of the following neighborhoods: Downtown, Midtown, Tech Square, and around such landmarks as The Fox Theater, Emory University Hospital Midtown, Mercedes-Benz Stadium, Home Depot Backyard, Centennial Olympic Park, Georgia Aquarium, World of Coca Cola, and parts of Renaissance Park.

Detroit:

In Detroit, 5G Ultra Wideband service will initially be concentrated in parts of the following areas: Dearborn, Livonia, and Troy, including areas around the Oakland-Troy Airport.

Source: Verizon


Update 1: Phoenix Launch + Boingo Partnership

Verizon’s 5G coverage is coming to Phoenix, AZ, bringing the list of 5G cities up to 10. The network will go live on August 23rd. Verizon also announced a partnership with Boingo to bring 5G Ultra Wideband service to indoor and public places.

This is important because Verizon’s current 5G network is essentially unusable indoors, a limitation of the technology they are using. The partnership should bring 5G to places like airports, stadiums, arenas, office buildings, hotels, etc.

Last, but not least, the Samsung Galaxy Note 10+ 5G will be available from Verizon tomorrow, August 23rd. The full retail price is $1,299.99.

Source: Verizon


Update 2: NYC, Boise, Panama City

Verizon’s 5G coverage is expanding to 3 more cities: New York City, Boise, and Panama City. In New York City, coverage will be in areas of Manhattan, Brooklyn, the Bronx, and around several landmarks. Verizon’s 5G technology limits coverage to very specific areas, so be sure to check the source below for all the exact locations you can access 5G in these cities.

Source: Verizon


Update 3: Omaha & Dallas

Today, Verizon has expanded 5G coverage to two more cities: Omaha, Nebraska and Dallas, Texas. This brings the number of cities with 5G coverage from Verizon up to 15. As with the previous announcements, the actual coverage areas are extremely specific. So if you live in these cities, be sure to check the link below to find out where you can get 5G speeds.

Source: Verizon


Update 4: Boston, Houston, and Sioux Falls

Verizon has announced that its 5G network is now live in three more cities across the US: Boston, MA, Houston, TX, and Sioux Falls, SD. This brings the total number of cities with Verizon 5G coverage up to 18. Just like the previous 15 cities, 5G is only accessible in these cities in very specific locations due to limitations with Verizon’s network technology. Be sure to visit the link below to see the exact locations where you can use 5G.

Source: Verizon


Update 5: 5G Coverage Map

Verizon has been flipping the switch for 5G in US cities for months, but they’ve never really had detailed coverage maps. You can now visit this page on Verizon’s website and select a city to see the 5G coverage. Maps show where 5G Ultra Wideband is strongest and you can zoom in to see LTE coverage as well. Verizon’s 5G coverage is very specific, so these maps are handy if you’re looking to try it out. The website also lists 10 cities that will get 5G next: Cincinnati, Kansas City, Charlotte, Little Rock, Cleveland, Memphis, Columbus, Salt Lake City, Des Moines, and San Diego.

Source: Verizon


Update 6: Los Angeles

Verizon 5G Ultra Wideband service is now available in areas around Los Angeles. As Verizon’s 5G network is limited to very specific locations, it’s not available city-wide. The exact locations are explained below, but Verizon will also have more detailed coverage maps available for the area on December 20th.

Parts of Downtown, Chinatown, Del Rey, and Venice around landmarks such as: Grand Park, Los Angeles Convention Center, Union Station, LA Live, Staples Center, and Venice Beach Boardwalk.

Source: Verizon

The post [Update 6: Los Angeles] Verizon 5G is Rolling Out to More Cities appeared first on xda-developers.



from xda-developers https://ift.tt/31aLIjV
via IFTTT

Samsung may soon introduce a new Samsung Flex program with Premium Care included

Samsung is a huge company with a lot of products and services to match. Currently, Samsung has Samsung Premium Care, which is an extended warranty program, and the Samsung Upgrade Program, which provides financing options and allows users to upgrade to the latest Samsung flagship every year. Now, Samsung is launching a new Samsung Flex program with Samsung Premium Care included.

An APK teardown can often predict features that may arrive in a future update of an application, but it is possible that any of the features we mention here may not make it in a future release. This is because these features are currently unimplemented in the live build and may be pulled at any time by the developers in a future build.

We were able to find references to the new Samsung Flex program in the latest Shop Samsung APK, version 1.0.16118. These strings state that Samsung Flex will let you upgrade to one of their products early. It will also include Samsung Premium Care.

<string name="samsung_flex">Samsung Flex</string>
<string name="samsung_flex_disclaimer">One monthly price. Premium Care support and service. Early upgrades to the latest %1$s. Cancel any time.*</string>

I think this could be an enhancement to the Samsung Upgrade program, but there is no way to be certain. Getting early access to devices does put a damper on that theory. This program will likely allow you to get some of their devices, with Premium Care included. As much info as these strings give us to what Premium Care is, it leaves us with more questions than it answers.

One thing I will say is if you are interested in getting this program for Samsung Premium Care, that’s a bad idea. Samsung Premium Care is one of the worst insurance programs I have ever personally used. Both XDA Portal Contributor Eric Hulse and I have recently tried to contact Asurion, the company that runs Premium Care. We have both experienced absolutely horrendous service. They go out of their way not to help us. They also have policies regarding the number of devices that can be covered under Premium Care, but they don’t tell you this unless you have an issue. Eric has, and pays for, 3 devices on Premium Care all under the same phone number. According to Asurion, you can only have 1 device per phone number. This isn’t publicly stated anywhere. He only found out about this policy after weeks of calls and messages with support.

While we don’t know much about this program, it does seem interesting. Paying a monthly fee for Premium Care and early access to upgrades on Samsung products seems like a great deal. Hopefully, we’ll learn more about this program soon. Just because it shows up in the strings, that doesn’t mean we will see it make the light of day, though. Hopefully we do, because these payment programs can be great for those who love to upgrade to the latest and greatest every year.


Thanks to PNF Software for providing us a license to use JEB Decompiler, a professional-grade reverse engineering tool for Android applications.

The post Samsung may soon introduce a new Samsung Flex program with Premium Care included appeared first on xda-developers.



from xda-developers https://ift.tt/2EpfPuR
via IFTTT

Vivo launches X30 and Vivo X30 Pro with 5G, Exynos 980 SoC, 64MP cameras, and FuntouchOS 10

Vivo might be alien to a lot of users in the west but it is fairly popular in parts of Asia. In fact, it is the second most popular smartphone brand in homeland China and occupies the third position in India. The company is also renowned for its design innovations like the first smartphone with a pop-up camera, multiple phones with AMOLED displays on both sides, or “waterfall displays” with >100% screen-to-body ratio. All these smartphones belong to the premium Vivo NEX series but today, they have launched two new smartphones in the X-series, which holds their affordable flagship smartphones. The successor to the Vivo X27 – Vivo X30 series – has just been launched in China and its exciting features include a 60x digital zoom, a minuscule hole-punch camera, 5G support, etc.

The camera is one of the primary features that Vivo focussed on while introducing the smartphone. The superior of the two – the Vivo X30 Pro – comes with a quad-camera setup on the back including a 64MP primary camera, a 13MP periscopic telephoto setup with up to 5X hybrid and 60X digital zoom, a 32MP camera with a 50mm focal length for depth sensing and better quality portraits, and lastly an 8MP wide-angle sensor. Additionally, there’s a LASER auto-focus on the camera module for faster and more accurate focusing.

On the other hand, the Vivo X30 comes with a triple-rear-camera setup but lacks a dedicated telephoto camera while retaining the remaining sensor, as they are. This means the zooming capabilities are limited to 20X on this smartphone but buyers will be able to capture DSLR-quality portrait shots, as claimed by the company, and capture a 112º wide field of view with the dedicated 8MP camera just like the Vivo X30 Pro.

Vivo announced that special camera features on the Vivo X30 and the X30 Pro include eye-tracking, Super Night Mode 2.0, and Hyper HDR for great shots irrespective of lighting conditions.

The Vivo X30 and the X30 Pro devices are identical in most other respects and in terms of the display and come equipped with a 6.44-inch Full HD+ Super AMOLED display. This display is being called an XDR display by Vivo for its claimed peak brightness of 1200 nits, HDR support, and a contrast ratio of 2000000:1. The display also houses a tiny, 2.98mm hole punch for the 32MP selfie camera. This hole-punch is the smallest we’ve seen so far and is less than 60% in diameter compared to the smallest hole-punch we’d seen before this.

When it comes to the performance, Vivo X30 and the X30 Pro are powered by Samsung’s Exynos 980 chipset. With this chipset, these smartphones also get support for NSA as well as SA modes for 5G. The maximum download speed for the modem on the Exynos 980 is 3.55Gbps.

The two smartphones come with 4350mAh batteries and support Vivo’s FlashCharge technology with a fast charging rate of 33W. To keep the phone from overheating due to demanding 5G connectivity, performance, or while fast charging, there is a gel-based cooling solution inside a flattened conduit.

The Vivo X30 series comes in three attractive colors – lilac-ish silver, peach, and black – all of which have a reflective gradient surface on the back.

In terms of software, both of the smartphones come with the latest version of Vivo’s custom Android skin – FuntouchOS. irrespective of the numbering, FuntouchOS 10 is based on Android 9 Pie but we can expect a version based on Android 10 when it launches for markets outside of China.

Vivo X30 5G and Vivo X30 Pro 5G: Pricing & Availability

Both the devices come with 8GB RAM as standard and options between 128GB and 256GB for the storage.

The prices for each of the models are:

  • Vivo X30
    • 8GB + 128GB – CNY 3298
    • 8GB + 256GB – CNY 3598
  • Vivo X30
    • 8GB + 128GB – CNY 3998
    • 8GB + 256GB – CNY 4298

The prebooking for the Vivo X30 starts today and the 8GB/128GB model of the Vivo X30 Pro goes on sale on December 24th. The 8GB/256GB variant of the X30 Pro and the 8GB/128GB variant of the X30 will be available December 28th onwards. Lastly, the launch date of the 8GB/256GB model of the X30 will be announced later.

FuntouchOS 10

FuntouchOS comes with a more minimal design include lighter backgrounds, fewer separating lines, and a coherent, pre-dominantly white color theme. We can also see improved symmetry in icons besides subtle animations. Along with these changes, there are new live wallpapers as well as animated always-on lock screen animations.

FuntouchOS 10 also bring some new gestures to view and quickly access notifications from favorite apps. Furthermore, there are several improvements to Vivo’s smart assistant Jovi. The interface will also be available for other Vivo devices starting Feb 2020.

funtouchos 10

The post Vivo launches X30 and Vivo X30 Pro with 5G, Exynos 980 SoC, 64MP cameras, and FuntouchOS 10 appeared first on xda-developers.



from xda-developers https://ift.tt/2rUNvOj
via IFTTT

Qualcomm Snapdragon 865 Benchmarks: Comparing CPU and GPU Performance with the Kirin 990, Snapdragon 855, and Snapdragon 845

Nearly two weeks ago, Qualcomm invited tech journalists to Maui for the 2019 Snapdragon Tech Summit. At the event, the company unveiled its latest high-end SoC for mobile devices: the Qualcomm Snapdragon 865 mobile platform. Qualcomm says the new Snapdragon 865 boasts a 25% CPU performance increase and a 20% GPU performance increase over the previous generation Snapdragon 855. Also, the new SoC supports LPDDR5 memory and is manufactured on a newer 7nm process. Qualcomm’s latest silicon will make its way to 2020 flagships like the Xiaomi Mi 10, OPPO Find X2, and many other high-end smartphones.

But just how much faster is it than the previous generations? We benchmarked Qualcomm’s Snapdragon 865 reference device at the event to find out. We pit the new SoC against the Snapdragon 855+, the Snapdragon 855, the Snapdragon 845, and the Kirin 990 from Huawei’s HiSilicon. We would have loved to test the Snapdragon 865 against the MediaTek Dimensity 1000 or Samsung Exynos 990, but sadly, there aren’t any devices with the new MediaTek and Samsung SoCs. Once we get our hands on real devices with the Snapdragon 865, we’ll be testing the real-world performance outside of benchmarks, too.


Qualcomm Snapdragon 865, Snapdragon 855, Snapdragon 845, and Kirin 990 Specifications

Qualcomm Snapdragon 865 Qualcomm Snapdragon 855+ Qualcomm Snapdragon 855 Qualcomm Snapdragon 845 HiSilicon Kirin 990 (4G)
CPU
  • 1 Kryo 585 ‘Prime’ (ARM Cortex-A77-based), up to 2.84GHz
  • 3 Kryo 585 ‘Performance’ (ARM Cortex-A77-based), up to 2.4GHz
  • 4 Kryo 385 ‘Efficiency’ (ARM Cortex-A55-based), up to 1.8GHz

25% Performance improvement over the previous generation

  • 1 Kryo 485 ‘Prime’ (ARM Cortex-A76-based), up to 2.96GHz
  • 3 Kryo 485 ‘Performance’ (ARM Cortex-A76-based), up to 2.42GHz
  • 4 Kryo 385 ‘Efficiency’ (ARM Cortex-A55-based), up to 1.8GHz
  • 1 Kryo 485 ‘Prime’ (ARM Cortex-A76-based), up to 2.84GHz
  • 3 Kryo 485 ‘Performance’ (ARM Cortex-A76-based), up to 2.42GHz
  • 4 Kryo 385 ‘Efficiency’ (ARM Cortex-A55-based), up to 1.8GHz

45% Performance improvement over the previous generation

  • 4 Kryo 385 ‘Performance’ (ARM Cortex-A75-based), up to 2.8GHz
  • 4 Kryo 385 ‘Efficiency’ (ARM Cortex-A55-based), up to 1.8GHz

25% Performance improvement over the previous generation

  • 2 ARM Cortex-A76, up to 2.86GHz
  • 2 ARM Cortex-A76, up to 2.09GHz
  • 4 ARM Cortex-A55, up to 1.86GHz
GPU Adreno 650

20% Performance improvement over the previous generation

Adreno 640 (15% overclocked) Adreno 640

20% Performance improvement over the previous generation

Adreno 630

25% Performance improvement over the previous generation

Mali-G76MP16
Memory 4x 16bit, 2133MHz LPDDR4X

4x 16bit, 2750MHz LPDDR5

4x 16bit, 2133MHz LPDDR4X 4x 16bit, 2133MHz LPDDR4X 4x 16-bit, 1866MHz LPDDR4X 4x 16-bit, LPDDR4X-4266
Manufacturing Process 7nm (TSMC N7P) 7nm (TSMC) 7nm (TSMC) 10nm LPP (Samsung) 7nm (TSMC)

Quick Overview of Each Benchmark

Benchmark explainer by Mario Serrafero

  • AnTuTu: This is a holistic benchmark. AnTuTu tests the CPU, GPU, and memory performance, while including both abstract tests and, as of late, relatable user experience simulations (for example, the subtest which involves scrolling through a ListView). The final score is weighted according to the designer’s considerations.
  • GeekBench: A CPU-centric test that uses several computational workloads including encryption, compression (text and images), rendering, physics simulations, computer vision, ray tracing, speech recognition, and convolutional neural network inference on images. The score breakdown gives specific metrics. The final score is weighted according to the designer’s considerations, placing a large emphasis on integer performance (65%), then float performance (30%) and finally crypto (5%).
  • GFXBench: Aims to simulate video game graphics rendering using the latest APIs. Lots of onscreen effects and high-quality textures. Newer tests use Vulkan while legacy tests use OpenGL ES 3.1. The outputs are frames during test and frames per second (the other number divided by the test length, essentially), instead of a weighted score.

    GFXBench Subscore Explanations. Click to expand.

    • Aztec Ruins: These tests are the most computationally heavy ones offered by GFXBench. Currently, top mobile chipsets cannot sustain 30 frames per second. Specifically, the test offers really high polygon count geometry, hardware tessellation, high-resolution textures, global illumination and plenty of shadow mapping, copious particle effects, as well as bloom and depth of field effects. Most of these techniques will stress the shader compute capabilities of the processor.
    • Manhattan ES 3.0/3.1: This test remains relevant given that modern games have already arrived at its proposed graphical fidelity and implement the same kinds of techniques. It features complex geometry employing multiple render targets, reflections (cubic maps), mesh rendering, many deferred lighting sources, as well as bloom and depth of field in a post-processing pass.
  • Speedometer, Jetstream: Javascript, core language features and performance on various operations; Javascript math, crypto, and search algorithm performance.
  • 3DMark (Sling Shot Extreme OpenGL ES 3.1/Vulkan): The test runs on a mobile-optimized rendering engine using OpenGL ES 3.1 and Vulkan (on Android) or Metal (on iOS). It comes with two subscores, each in turn featuring multiple subscores, all of which ultimately use frames per second as their metric across multiple testing scenarios. This benchmark will test the full range of API features, including transform feedback, multiple render targets and instanced rendering, uniform buffers, and features such as particle illumination, volumetric lighting, deferred lighting, depth of field and bloom in post-processing, all using compute shaders. Offscreen tests use a fixed time step between frames, and rule out any impact caused by vertical sync, display resolution scaling and related OS parameters. The final score is weighted according to the designer’s considerations.3DMark score weights
  • PCMark 2.0:  Tests the device as a complete unit. It simulates everyday use cases that can implement abstract algorithms and a lot of arithmetic; the difference is that these are dispatched within an application environment, with a particular practical purpose, and handled by API calls and Android libraries common to multiple applications. The test will output a variety of scores corresponding to the various subtests, which will be detailed below; the composite, Work 2.0 score is simply the geometric mean of all of these scores, meaning all tests are weighted equally.

    PCMark 2.0 Subscore Explanations. Click to expand.

    • Web browsing 2.0 simulates browsing social media: rendering the web page, searching for the content, re-rendering the page as new images are added, and so on. This subtest uses the native Android WebView to render (WebKit) and interact with the content, which is locally stored — this means you can run it offline, but it does not simulate web browsing fully as it rules out internet connection factors (latency, network speed). It is specifically tracking frame rates and completion time across seven tasks, with their score being a multiple of their geometric mean.
    • Video Editing simulates video editing performance: applying effects to a video using OpenGL ES 2.0 fragment shaders, decoding video frames (sent to an Android GLSurfaceView), and rendering/encoding the video in H.264/MPEG-4AVC at several frame rates and resolutions up to 4K. It is specifically tracking frame rates on the UI, except for a final test tracking the completion time of a video editing pipeline.
    • Writing simulates general document and text editing work: adding or editing texts and images within a document, copying and pasting text, and so on. It uses the native Android EditText view as well as PdfRenderer and PdfDocument APIs. It will open compressed documents, move text bodies, insert images in the document, then save them as a PDF, to then encrypt and decrypt them (AES). It specifically tracks task completion times for the processes of opening and saving files, adding images and moving text bodies, encrypt/decrypt the file, and render the PDF pages on ImageViews.
    • Photo Editing simulates photo-editing performance: opening images, applying different effects via filters (grains, blurs, embossing, sharpening and so on) and saving the image. It uses 4MP JPEG source images and manipulates them in bitmap format using the android.media.effect API, android.renderscript API’s RenderScript Intrinsics, android-jhlabs, and the native android.graphics API for drawing the process on the screen. This is an extremely comprehensive test in that it will be impacted by storage access, CPU performance, GPU performance, and it is dependent on many different Android APIs.  The test specifically measures memory and storage access times, encoding and decoding times, task completion times. The various filters and effects come from different APIs.
    • Data manipulation simulates database management operations: parsing and validating data from files, interacting with charts, and so on. It will open (date, value) tuples from CSV, XML, JSON files and then render animated charts with the MPAndroidChart library. It specifically tracks data parsing times as well as draws per second of each chart animation (similar to frame rate, but specific to the updating chart).

Source links for each benchmark can be found at the end of the article.


Test Devices

Qualcomm Snapdragon 865 Qualcomm Snapdragon 855+ Qualcomm Snapdragon 855 Qualcomm Snapdragon 845 HiSilicon Kirin 990
Device Name Qualcomm Reference Device (QRD) ASUS ROG Phone II Google Pixel 4 Google Pixel 3 XL Huawei Mate 30 Pro
Software Android 10 (Qualcomm customized AOSP software) Android 9 (ZenUI 6.0 OEM software with October 2019 security patch) Android 10 (Google Pixel OEM software with December 2019 security patch) Android 10 (Google Pixel OEM software with December 2019 security patch) Android 10 (EMUI 10.0 OEM software with October 2019 security patch)
Display 2880×1440 @ 60Hz 2340×1080 @ 60Hz 2280×1080 @ 60Hz 2960×1440 @ 60Hz 2400×1176 @ 60Hz
Memory 12GB LPDDR5 8GB LPDDR4X 6GB LPDDR4X 4GB LPDDR4X 8GB LPDDR4X
Storage 128GB UFS 3.0 128GB UFS 3.0 64GB UFS 2.1 64GB UFS 2.1 256GB UFS 3.0
Performance Mode Yes* No No No No

*Performance mode on the Snapdragon 865 QRD makes workloads appear 20% “heavier” to the scheduler. This means that a CPU that is loaded 80% will appear 100% loaded to the scheduler, ramping up clocks faster and migrating tasks from the little to the big cores faster. However, CPU clock speeds are NOT boosted.


Benchmark Results

Main Scores

Benchmark Version Qualcomm Snapdragon 865 Qualcomm Snapdragon 855+ Qualcomm Snapdragon 855 Qualcomm Snapdragon 845 HiSilicon Kirin 990
AnTuTu 8.0.4 565,384 425,963 386,499 278,647 389,505
Geekbench single-core 5.0.2 929 760 600 521 750
Geekbench multi-core 5.0.2 3,450 2,840 2,499 2,125 2,887
GFXBench ES 3.0 1080 Manhattan offscreen 5.00 126 110 92 82 104
GFXBench ES 3.1 1080 Carchase offscreen 5.00 50 48 40 35 38
GFXBench ES 3.1 1080 Manhattan offscreen 5.00 88 78 67 61 67
GFXBench ES 2.0 1080 T-Rex offscreen 5.00 205 185 164 152 105
GFXBench 1440p Aztec Ruins Vulkan (High Tier) Offscreen IFH 5.00 20 19 16 14 16
GFXBench 1440p Aztec Ruins OpenGL (High Tier) Offscreen IFH 5.00 20 18 16 14 18
Speedometer 2.00 80 36 53 49 65.4
JetStream – Geometric mean 1.10 123 116 98 85 95.8
PCMark – Work 2.0 2.0.3716 12,626 9,068 9,311 8,988 8,667
Androbench Sequential Read (MB/s) 5.0.1 1,459 1,398 873 659 1,451.09
Androbench Sequential Write (MB/s) 5.0.1 225 217 189 231 443.66
Androbench Random Read (IOPS) 5.0.1 50,378 41,315 37,600 32,376 53,114.78
Androbench Random Write (IOPS) 5.0.1 48,410 35,422 41,340 37,417 55,972.18
Androbench Random Read (MB/s) 5.0.1 195 161 147 126 207.47
Androbench Random Write (MB/s) 5.0.1 189 138 161 146 218.64
Androbench SQLite Insert 5.0.1 3,705 3,187 3,207 2,627 4,968.81
Androbench SQLite Update 5.0.1 4,014 3,931 3,996 3,333 6,090.65
Androbench SQLite Delete 5.0.1 5,037 4,964 4,558 4,081 7,664.88
3DMark Sling Shot Extreme Open GL ES 3.1 Overall Score 2.0.4646 7,008 6,201 5,174 3,431 5,677
3DMark Sling Shot Extreme Vulkan Overall Score 2.0.4646 6,449 5,339 4,339 3,273 4,303

Subscores

Benchmark Subscore Chart. Click to expand.

Benchmark Subscore Qualcomm Snapdragon 865 Qualcomm Snapdragon 855+ Qualcomm Snapdragon 855 Qualcomm Snapdragon 845
AnTuTu CPU 182,101 118,473 117,500 77,245
CPU Mathematical Operations 47,555 33,101 35,852 19,449
CPU Common Algorithms 40,260 23,468 20,400 13,203
CPU Multi-Core 94,286 61,904 61,248 44,593
GPU 218,496 193,905 160,291 117,022
GPU Terracotta – Vulkan 54,634 49,080 40,874 33,176
GPU Coastline – Vulkan 77,022 68,847 49,274 36,549
GPU Refinery – OpenGL ES3.1+AEP 86,840 75,978 70,143 58,356
MEM 81,392 65,011 56,889 46,041
MEM RAM Access 37,450 27,154 25,031 19,153
MEM ROM App IO 4,876 4,785 4,914 4,539
MEM ROM Sequential Read 22,039 20,046 13,240 9,499
MEM ROM Sequential Write 3,513 3,309 2,891 3,328
MEM ROM Random Access 13,514 9,718 10,813 9,523
UX 83,396 48,573 51,818 38,339
UX Data Security 13,788 8,835 9,384 6,041
UX Data Processing 28,615 9,852 9,088 5,959
UX Image Processing 14,473 9,799 12,741 10,192
UX User Experience 26,520 20,088 20,605 16,147
3DMark Sling Shot Extreme Open GL ES 3.1 Graphics Score 8,158 7,092 5,631 3,384
Sling Shot Extreme Open GL ES 3.1 Physics Score 4,693 4,308 4,401 3,623
Sling Shot Extreme Vulkan Graphics Score 8,224 6,557 4,845 3,425
Sling Shot Extreme Vulkan Physics Score 3,674 3,246 3,177 2,835
PCMark Web Browsing 2.0 score 11,680 6,427 6,985 7,806
Video Editing score 6,575 5,894 5,611 6,638
Writing 2.0 score 14,389 11,475 10,945 9,364
Photo Editing 2.0 score 36,868 18,247 22,159 17,516
Data Manipulation score 7,880 7,732 7,361 6,902
Geekbench Single-core Crypto Score 1,435 1,055 873 838
Single-core Integer Score 878 736 578 513
Single-core Floating Point Score 956 762 604 488
Multi-core Crypto Score 5,594 3,874 3,746 3,703
Multi-core Integer Score 3,304 2,764 2,410 2,093
Multi-core Floating Point Score 3,412 2,831 2,482 1,930

Main Scores Comparison

Subscore Versus Snapdragon 865 Versus Snapdragon 855+ Versus Snapdragon 855 Versus Snapdragon 845 Versus Kirin 990
AnTuTu 1x 1.33x 1.46x 2.03x 1.45x
Geekbench single-core 1x 1.22x 1.55x 1.78x 1.24x
Geekbench multi-core 1x 1.21x 1.38x 1.62x 1.2x
GFXBench ES 3.0 1080 Manhattan offscreen 1x 1.15x 1.37x 1.54x 1.21x
GFXBench ES 3.1 1080 Carchase offscreen 1x 1.04x 1.25x 1.43x 1.32x
GFXBench ES 3.1 1080 Manhattan offscreen 1x 1.13x 1.31x 1.44x 1.31x
GFXBench ES 2.0 1080 T-Rex offscreen 1x 1.11x 1.25x 1.35x 1.95x
GFXBench 1440p Aztec Ruins Vulkan (High Tier) Offscreen IFH 1x 1.05x 1.25x 1.43x 1.25x
GFXBench 1440p Aztec Ruins OpenGL (High Tier) Offscreen IFH 1x 1.11x 1.25x 1.43x 1.11x
Speedometer 1x 2.22x 1.51x 1.63x 1.22x
JetStream – Geometric mean 1x 1.06x 1.26x 1.45x 1.28x
PCMark – Work 2.0 1x 1.39x 1.36x 1.4x 1.46x
Androbench Sequential Read (MB/s) 1x 1.04x 1.67x 2.21x 1.01x
Androbench Sequential Write (MB/s) 1x 1.04x 1.19x 0.97x 0.51x
Androbench Random Read (IOPS) 1x 1.22x 1.34x 1.56x 0.95x
Androbench Random Write (IOPS) 1x 1.37x 1.17x 1.29x 0.86x
Androbench Random Read (MB/s) 1x 1.21x 1.33x 1.55x 0.94x
Androbench Random Write (MB/s) 1x 1.37x 1.17x 1.29x 0.86x
Androbench SQLite Insert 1x 1.16x 1.16x 1.41x 0.75x
Androbench SQLite Update 1x 1.02x 1x 1.2x 0.66x
Androbench SQLite Delete 1x 1.01x 1.11x 1.23x 0.66x
3DMark Sling Shot Extreme Open GL ES 3.1 Overall Score 1x 1.13x 1.35x 2.04x 1.23x
3DMark Sling Shot Extreme Vulkan Overall Score 1x 1.21x 1.49x 1.97x 1.50x

Subscores Comparison

Benchmark Subscores Comparison Chart. Click to expand.

Benchmark Subscore Versus Snapdragon 865 Versus Snapdragon 855+ Versus Snapdragon 855 Versus Snapdragon 845
AnTuTu CPU 1x 1.54x 1.55x 2.36x
CPU Mathematical Operations 1x 1.44x 1.33x 2.45x
CPU Common Algorithms 1x 1.72x 1.97x 3.05x
CPU Multi-Core 1x 1.52x 1.54x 2.11x
GPU 1x 1.13x 1.36x 1.87x
GPU Terracotta – Vulkan 1x 1.11x 1.34x 1.65x
GPU Coastline – Vulkan 1x 1.12x 1.56x 2.11x
GPU Refinery – OpenGL ES3.1+AEP 1x 1.14x 1.24x 1.49x
MEM 1x 1.25x 1.43x 1.77x
MEM RAM Access 1x 1.38x 1.5x 1.96x
MEM ROM App IO 1x 1.02x 0.99x 1.07x
MEM ROM Sequential Read 1x 1.1x 1.66x 2.32x
MEM ROM Sequential Write 1x 1.06x 1.22x 1.06x
MEM ROM Random Access 1x 1.39x 1.25x 1.42x
UX 1x 1.72x 1.61x 2.18x
UX Data Security 1x 1.56x 1.47x 2.28x
UX Data Processing 1x 2.9x 3.15x 4.8x
UX Image Processing 1x 1.48x 1.14x 1.42x
UX User Experience 1x 1.32x 1.29x 1.64x
3DMark Sling Shot Extreme Open GL ES 3.1 Graphics Score 1x 1.15x 1.45x 2.41x
Sling Shot Extreme Open GL ES 3.1 Physics Score 1x 1.09x 1.07x 1.3x
Sling Shot Extreme Vulkan Graphics Score 1x 1.25x 1.7x 2.4x
Sling Shot Extreme Vulkan Physics Score 1x 1.13x 1.16x 1.3x
PCMark Web Browsing 2.0 score 1x 1.82x 1.67x 1.5x
Video Editing score 1x 1.12x 1.17x 0.99x
Writing 2.0 score 1x 1.25x 1.31x 1.54x
Photo Editing 2.0 score 1x 2.02x 1.66x 2.1x
Data Manipulation score 1x 1.02x 1.07x 1.14x
Geekbench Single-core Crypto Score 1x 1.36x 1.64x 1.71x
Single-core Integer Score 1x 1.19x 1.52x 1.71x
Single-core Floating Point Score 1x 1.25x 1.58x 1.96x
Multi-core Crypto Score 1x 1.44x 1.49x 1.51x
Multi-core Integer Score 1x 1.2x 1.37x 1.58x
Multi-core Floating Point Score 1x 1.21x 1.37x 1.77x

Concluding Highlights

Analysis by Mario Serrafero:

  • For AnTuTu’s final score, we observe a large 33% bump over the 855+ and a massive improvement of around 45% over the 855. The CPU subtests showcase massive improvements, with uplifts in each subscore ranging from 15% to 97%. These results are surprising given that Qualcomm posted a respectable 25% CPU performance uplift over the Snapdragon 855, yet we see all CPU subscores go up by over 40%, and even 70%. The GPU side of the subscores, however, sees a much more restrained increase of around 13% on average, compared to the 855+, or 24% to 56% compared to our 855 scores from the Google Pixel 4.
  • The popular PCMark 2.0 saw a massive jump of almost 40% in its “Work 2.0” final score, compared to the 855+. Looking at the subscores, it seems that most of the improvement lies in the Photo Editing 2.0 subtest, which nearly doubles in score, followed by a Web Browsing score improvement of around 80%. The final score is simply the average between all subscores, so these massive bumps end up being balancing out the more conservative figures of the other subscores, which remain constant or rise by less than 25%.
  • Geekbench 5 subscores gave us a decent look into where the resulting ~20% increase in Single-core and Multi-core scores comes from. The crypto tests (which are weighted the least in calculating the final scores) had a performance increment of 36% and 44% (single and multi, respectively) compared to our 855+ results, whereas integer and floating-point performance only rose by about 19% to 25%, perfectly in-line with Qualcomm’s figures. The gap is much larger if we compare the 865 to our 855 results from the Pixel 4, as crypto goes up by 66% while integer and floating-point improvements sit over 50% for single-core tests and over 35% for multi-core tests. Given the 865 features the same clock speeds as the 855, we see a bump in integer and floating score performance per MHz.
  • 3DMark scores also fall more-or-less in line with the expected 20% faster graphics rendering that Qualcomm boasted at the Snapdragon tech summit. The graphics and physics scores saw an increase of 15% and 11% (respectively) over the 855+ for the OpenGL ES 3.1 test, and 25% and 22% for the Vulkan test. This suggests the 865 is a healthy upgrade for gamers.
  • GFXBench only saw a performance boost of 5% to 15% over the 855+, though when comparing it against the regular 855 those numbers jump above the 20% year-on-year increments posted by the company.

Recommended Reading


Benchmark Sources

CPU, GPU, and Memory

AnTuTu Benchmark (Free, Google Play) →

CPU and Memory

Geekbench 5 (Free, Google Play) →

System

PCMark for Android Benchmark (Free, Google Play) →

GPU

GFXBench Benchmark (Free, Google Play) →

3DMark - The Gamer's Benchmark (Free, Google Play) →

Storage

Androbench (Storage Benchmark) (Free, Google Play) →

Browser

Speedometer 2.0 ||| JetStream 1.1


Thanks to TK Bay for the featured image. Thanks to Max Weinbach for providing the Kirin 990 results from his Huawei Mate 30 Pro.

The post Qualcomm Snapdragon 865 Benchmarks: Comparing CPU and GPU Performance with the Kirin 990, Snapdragon 855, and Snapdragon 845 appeared first on xda-developers.



from xda-developers https://ift.tt/2Pp2kSe
via IFTTT