A shadow mapping test:
This section allows you to view all posts made by this member. Note that you can only see posts made in areas you currently have access to.
During today's Epic Games event at the Game Developers Conference 2015, NVIDIA co-founder Jen-Hsun Huang rushed the stages like a professional wrestling hero to announce the Titan X, NVIDIA's latest GPU. Huang claims it is the most powerful GPU on the planet. With 12GB frame buffer and 8 billion transistors, it is — on paper — a significant step past NVIDIA's current hardware. NVIDIA's Titan used to be its most powerful hardware.
As there is no compatibility between OpenGL and the Vulkan API, many large GL code bases will not switch immediately. But older applications will still want to access new hardware features, so I believe that OpenGL is not dead yet, but will in fact live on for some more versions. I’m confident that we will see new hardware features exposed in OpenGL 5.0 (maybe at Siggraph this year?). Just as NVIDIA, AMD and Intel didn’t drop the Compatibility profile even in the latest OpenGL versions, I don’t expect them to drop support for new OpenGL versions any time soon. I guess it’s actually more likely that we will quite some OpenGL extensions which expose more Vulkan like features in OpenGL to help transitioning (e.g. the Vulkan shader “binaries” SPIR-V). Limited compatibility in the other direction at least for the shaders is planned by providing a GLSL to SPIR-V compiler.
Imagination is a promoting member of the Khronos Group and has been working on developing a proof-of-concept driver for Vulkan for our PowerVR Rogue GPUs. Our PowerVR demo team has also spent the last two months porting one of our new OpenGL ES 3.0 demos to the new API and today we are able to show you a snapshot of our work.
For example, there are no glUniform*() equivalent entry points in Vulkan; instead, writing to GPU memory is the only way to pass data to shaders.
Command buffers can be created on a different thread to the thread they are submitted on. This means rendering commands could be created on all cores of a CPU.
When you call glTexStorage2D() in OpenGL, the driver has to allocate memory for a two-dimensional or one-dimensional array texture. The function and the memory allocation process represent a black box.
In Vulkan however, the memory allocation is done by the application. This means that the application knows more about what type of memory it is using and more importantly how much memory it is using, which should be useful for applications that are memory-bound.
SPIR-V is a new platform-independent intermediate language. It is a self-contained, fully specified, binary format
for representing graphical-shader stages and compute kernels for multiple APIs. Physically, it is a stream of 32-bit words. Logically, it is a header and a linear stream of instructions. These encode, first, a set of annotations and decorations, and second a collection of functions. Each function encodes a control-flow graph (CFG) of basic blocks, with additional instructions to preserve source-code structured flow control. Load/store instructions are used to access declared variables, which includes all input/output (IO). Intermediate results bypassing load/store use single static-assignment (SSA) representation. Data objects are represented logically, with hierarchical type information: There is no flattening of aggregates or assignment to physical register banks, etc. Selectable addressing models establish whether general pointers may be used, or if memory access is purely logical.
AMD is a company that fundamentally believes in technologies unfettered by restrictive contracts, licensing fees, vendor lock-ins or other arbitrary hurdles to solving the big challenges in graphics and computing. Mantle was destined to follow suit, and it does so today as we proudly announce that the 450-page programming guide and API reference for Mantle will be available this month (March, 2015) at www.amd.com/mantle
Mantle’s definition of “open” must widen. It already has, in fact. This vital effort has replaced our intention to release a public Mantle SDK, and you will learn the facts on Thursday, March 5 at GDC 2015.
Unreal Engine 4 is now available to everyone for free, and all future updates will be free!
You can download the engine and use it for everything from game development, education, architecture, and visualization to VR, film and animation. When you ship a game or application, you pay a 5% royalty on gross revenue after the first $3,000 per product, per quarter. It’s a simple arrangement in which we succeed only when you succeed.
This is the complete technology we use at Epic when building our own games. It scales from indie projects to high-end blockbusters; it supports all the major platforms; and it includes 100% of the C++ source code. Our goal is to give you absolutely everything, so that you can do anything and be in control of your schedule and your destiny. Whatever you require to build and ship your game, you can find it in UE4, source it in the Marketplace, or build it yourself – and then share it with others.
Companies from Facebook Inc. to Sony Corp. and Google Inc. have spent billions of dollars investing in virtual-reality technology.
Now they have to turn it into a business. To get consumers to buy their devices, Sony, Facebook’s Oculus VR, Valve Corp. and Razer USA Ltd. need games to play and videos to watch. At this week’s Game Developers Conference in San Francisco, they’ll be trying to convince software makers there’s a market to be had.
“The market for virtual reality is very hot right now, and a lot of highly visible developers, designers and investors are placing bets,” van Dreunen said. “However, in the absence of solid delivery dates from the leading companies in the space, it is clear that virtual reality is not yet ready for prime time.”
Rays are fired from the camera out into the scene, bounce around according to physical models, and record when they hit a light source. Whenever anything changes, the scene has to be re-traced (it starts out noisy, but converges to the correct image).
Google's search engine currently uses the number of incoming links to a web page as a proxy for quality, determining where it appears in search results. So pages that many other sites link to are ranked higher. This system has brought us the search engine as we know it today, but the downside is that websites full of misinformation can rise up the rankings, if enough people link to them.
A Google research team is adapting that model to measure the trustworthiness of a page, rather than its reputation across the web. Instead of counting incoming links, the system – which is not yet live – counts the number of incorrect facts within a page.
Here’s a short summary of the reasons:
- “What I do is not interesting to others”
- Coding is better than writing
- I’m not doing anything innovative
- “I don’t do side projects”
- “I don’t know what to write about”
- “what I want to write about is too obvious”
- writing is harder than coding
- everyone knows that already
- longer feedback loop, as compared to coding
or quite some time now, you’ve been hearing about Qt3D’s Framegraph. Although a brief definition of what the Framegraph is was given in the previous articles, this blog post will cover it in more detail. After reading this post, you will understand the difference between the Scenegraph and the Framegraph and see their respective uses. The more adventurous amongst you, will be able to pick up a pre-release version of Qt3D and start experimenting to see what the Framegraph can do for you.