Anteru's blog
  • Consulting
  • Research
    • Assisted environment probe placement
    • Assisted texture assignment
    • Edge-Friend: Fast and Deterministic Catmull-Clark Subdivision Surfaces
    • Error Metrics for Smart Image Refinement
    • High-Quality Shadows for Streaming Terrain Rendering
    • Hybrid Sample-based Surface Rendering
    • Interactive rendering of Giga-Particle Fluid Simulations
    • Quantitative Analysis of Voxel Raytracing Acceleration Structures
    • Real-time Hybrid Hair Rendering
    • Real-Time Procedural Generation with GPU Work Graphs
    • Scalable rendering for very large meshes
    • Spatiotemporal Variance-Guided Filtering for Motion Blur
    • Subpixel Reconstruction Antialiasing
    • Tiled light trees
    • Towards Practical Meshlet Compression
  • About
  • Archive

What's going on?

June 05, 2007
  • General
  • Programming
approximately 1 minutes to read

Quite some time since my last post - I’ve been rather really busy will different stuff. First of all, I had to work a bit more as I was falling behind the schedule, but I’m on track now again. I didn’t work much on niven lately, but the new object management is in good shape. I polished the integrated math suite a bit, it’s more flexible and generic than before and has a few more useful features. For university, I took some time to implement a few things - iterative solvers, FFT and LUP-Decomposition. Read on for more details.

  • Iterative solvers - solve equation systems by iterating from a first (guessed) solution. Sorted by difficulty … actually, Richardson is the most tricky, as you need a correction factor smaller than twice the spectral radius of the matrix (if you think hu? - I had also no idea, but it turns out that the spectral radius is less or equal to any matrix norm, so it’s not that difficult) - Richardson - Jacobi - Gauss-Seidel, normal and with successive over-relaxation (which did not change much on the test data)
  • FFT - fast polynomial multiplication by transforming between coefficient and point-value-representation. Although I did a “by-the-book” implementation I’m not satisfied at all with the quality - the Intel MKL offers much better numeric stability than the naive solution I implemented. I think this might come from the following issue: Instead of computing the n-th root directly, I’m starting from the first one and multiplying it n-times; I assume a good complex exponentiation to be much more precise here.
  • LUP-Decomposition - fast and exact solving of linear equations, plus very fast determinants. I implemented it with pivoting so I can solve a wider range of problems and I get better numeric stability. This one was real fun, and it turned out to be simpler than expected.
Previous post
Next post

Recent posts

  • Data formats: Why CSV and JSON aren't the best
    Posted on 2024-12-29
  • Replacing cron with systemd-timers
    Posted on 2024-04-21
  • Open Source Maintenance
    Posted on 2024-04-02
  • Angular, Caddy, Gunicorn and Django
    Posted on 2023-10-21
  • Effective meetings
    Posted on 2022-09-12
  • Older posts

Find me on the web

  • GitHub
  • GPU database
  • Projects

Follow me

Anteru NIV_Anteru
Contents © 2005-2025
Anteru
Imprint/Impressum
Privacy policy/Datenschutz
Made with Liara
Last updated February 03, 2019