00:00 | Hi my name is Peru and I'm to introduce you to my research in |
|
00:04 | mini talk. Let's get started My overall research interest is in the |
|
00:11 | , high performance computing HPC area which a victim that says something like computing |
|
00:18 | performance matters a lot mostly. This heavy numerical computations such as simulations of |
|
00:24 | chemical biological social phenomenon at a very scale and with very high fatality. |
|
00:30 | of my main research topic is in extension computations. And the goal is |
|
00:36 | make it fast scalable and energy efficient make use of the latest advancements in |
|
00:43 | hardware such as Gpus clusters, clouds special units such as tension processing |
|
00:48 | And potential core also have some other in algorithms, especially in numerical algorithms |
|
00:55 | computer systems in complete scientific software and a visualization as you can see, |
|
01:03 | research is at the intersection of numerical , HPC software system and libraries and |
|
01:10 | of computer hardware. We investigate the of the of the advancements in these |
|
01:16 | areas. Let me introduce a few snapshots. So the first snapshot is |
|
01:23 | project that I call later short for algebra on terms of course. So |
|
01:28 | is tensile court basically in terms of is a device on your on NVIDIA |
|
01:33 | that are designed to accelerate deep neural . Um it comes it comes with |
|
01:39 | video GPU that appears since 2018. big deal about tensile core is that |
|
01:45 | specializes in computing Matrix matrix multiplication at very high speed hundreds of teraflops. |
|
01:53 | one teraflops means 10 to the power 12 numerical operations per second. So |
|
02:00 | is equivalent to about 1000 CPU cores the same thing at the same |
|
02:06 | They are designed for neural networks. , we'd like to make use of |
|
02:10 | to do more than that to compute sentence for factories. Asians, for |
|
02:15 | , this open doors too much wider in scientific computing engineering and data science |
|
02:21 | , it turns out that to do we need some innovation in numerical |
|
02:26 | software and performance engineering and basically a level concerted algorithmic and performance oriented |
|
02:35 | We are able to achieve 3 to times speed up for major Matrix Factory |
|
02:41 | called Q. R. Factory This is a significant improvement with the |
|
02:45 | of tensile court. This kind of operations are very useful and they are |
|
02:50 | the standard two blocks of every major processor vendors and they are generally highly |
|
02:58 | at better companies because they are the blocks for many, many applications. |
|
03:04 | second research snapshot is on matrix or decomposition and a competition. These are |
|
03:10 | diverse and useful for many problems in learning, signal processing, bioinformatics and |
|
03:17 | areas. For example, in data , a powerful unsupervised model for topics |
|
03:22 | of documents. Gaussian mixture model, recognition recommendation systems and hyper spectral and |
|
03:30 | just maybe a few examples. The is that it's very time consuming and |
|
03:36 | requires a lot of memory to compute scale major extensive decomposition or competition. |
|
03:44 | , so how do we scale and this particular operation? So we do |
|
03:49 | in parallel in distributed memory system with help of GPU and we also borrow |
|
03:55 | innovate uh, the algorithms such as and communication, avoiding in algebra and |
|
04:05 | order optimization algorithms. And we work high performance platforms with M P |
|
04:11 | code A and C plus plus. third snapshot is on the class of |
|
04:18 | models called kernel machines for classification and . This include support vector machine and |
|
04:26 | model and Gaussian process. When you more points, the model fits the |
|
04:32 | better and also reduces uncertainty at the that he has never seen before with |
|
04:38 | two models. The challenges again that becomes quite challenging at large scale. |
|
04:44 | we also try to solve this problem HPC. These are the most relevant |
|
04:50 | for my research and thank you very for listening. |
|