Linley Spring Processor Conference 2020
Held April 6-9, 2020
Proceedings available
» Events | Event Info | Day One | Day Three | Day Four | Proceedings & Event Videos
Agenda for Day Two: Tuesday April 7, 2020
View Day One
9:00am-12:30pm | Accelerating AI and Other Embedded Workloads
New CPUs and accelerators enable embedded processors to handle diverse workloads, such as on-device AI, high-speed networking, sensor signal processing, and 5G connectivity. To optimize performance, ASIC designers need configurable IP cores they can tune to their system requirements. Regardless of the target market, users need tools that ease software development. This session, led by The Linley Group senior analyst Mike Demler, will cover chips and licensable cores, as well as software tools that accelerate embedded applications. |
9:00am-9:20am | SiMa.ai
Machine learning (ML) at the edge must deliver cloud-like performance within stringent power envelopes. When operating autonomous automobiles, drones, surveillance cameras, and robots, a typical CPU can't meet the ML performance requirements for many new systems, thus requiring an ML accelerator. This presentation will illustrate how the industry's first green machine-learning SoC (MLSoC) provides custom ML acceleration that delivers more than 30x improvement in frames per second per watt (FPS/W) over alternatives.
|
9:20am-9:30am | Question and Answer |
9:30am-9:50am | SiFive
Peak FLOPs and observed FLOPS have little correlation in vector architectures. Peak FLOPs are meaningless when the architecture is stalled on memory. The key to "achieving supercomputer performance", as Dongarra has called it, is to combine a vector unit with "real" vector registers and an intelligent restructuring compiler to focus computations into the processor. This presentation will discuss SiFive's RISC-V Vector extension and the technical tradeoffs among the three components, illustrating the resulting advantages with specific benchmark performance.
|
9:50am-10:00am | Question and Answer |
10:00am-10:20am | Flex Logix
eFPGA technology offers functionality similar to leading FPGA chips and is proven from 7nm to 180nm. But eFPGA designs can be iterated in less than a year, and architectures can evolve rapidly. Unlike standard FPGAs that have a bit of everything, eFPGA can focus on the acceleration functions that matter for a given application. This presentation will describe how architectures are innovating to accelerate workloads like floating-point signal processing and AI inference in both INT8 and BF16 formats.
|
10:20am-10:30am | Question and Answer |
10:30am-10:40am | Break Sponsored by Flex Logix |
10:40am-11:00am | Cadence
Smartphones, digital assistants, cars, and other embedded systems are using machine learning for audio and speech applications. Typically, neural-network models are hand coded to fit within the memory and power constraints of the system. The latest capabilities of TensorFlow can provide an end-to-end framework enabling users to train and efficiently deploy their models. This presentation will walk through the details of building a keyword-detection model and deploying it on a HiFi DSP using the TensorFlow tool chain.
|
11:00am-11:10am | Question and Answer |
11:10am-11:30am | CEVA
Contextual awareness is enabling enriched user experience in applications such as robotics, wearables, and conversational assistants. It involves processing of multiple sensors such as cameras, radars, microphones, and motion sensors, and includes various computer vision, voice processing, and neural network workloads. This presentation will introduce a new scalable vector DSP architecture for contextual awareness processing, providing floating-point arithmetic, optimized linear-algebra operations, and an enhanced pipeline for higher quality and power efficiency.
|
11:30am-11:40am | Question and Answer |
11:40am-12:00pm | Synopsys
A transformation is occurring in high-end embedded applications as Internet infrastructure spreads from the cloud to the edge. Performance requirements are increasing, but designers must account for power and area limits. These changes are driving greater use of configurable multicore solutions and specialized hardware accelerators. This presentation will introduce a new multiprocessor cluster architecture that is highly scalable, enabling it to achieve very high performance for storage, automotive, networking and other high-end embedded applications.
|
12:00pm-12:10pm | Question and Answer |
12:10pm-1:10pm | Breakout sessions with today's speakers |