If this is your first visit, be sure to
check out the FAQ by clicking the
link above. You may have to register
before you can post: click the register link above to proceed. To start viewing messages,
select the forum that you want to visit from the selection below.
In that case, I didn't know what you meant. I thought the situation was something like, "we now know how to correctly program these chips within the OSS design". Instead, it seems that you were actually missing hardware details.
It's a bit of both. In theory we have all the hardware details; in practice we have the hardware details from when the hardware was designed, which is theoretically the same as what was implemented but in practice may be a bit different. The last couple of generations haven't had that problem but it was a real challenge back in the 5xx/6xx days. We can also comb through bug reports and the associated fixes, but anyone who has searched an engineering-level bug tracker for matches to their symptoms understands how fruitless that can be despite everyone's best efforts.
In many cases we know what registers need to be programmed and what to put in them, but there are subtle seauencing dependencies which are not documented anywhere.
The devs have fairly easy access to 99% of the HW details and can work through the driver and hardware teams to get to maybe 99.9% fairly easily. That last 0.1% is a real pain though... and quite often the chip seems completely broken until you find that last 0.1% ;(
There are all kinds of troubleshooting tools, it's just that modern high end GPUs are really big and complex. That works OK when the effort is spread across thousands of people, but it's a bit hard for a small team to do the same. That's why I have hope for piggybacking on the original design effort.
Now, this comment obviously comes from somebody whose coding skills go no further than solving equations in python, but, if some OSS developers have access to the millions of lines of code you talk about, i.e. they work not only from the released documentation but also from closed code, how do they manage to restrict themselves and not implement the solutions adopted in the closed driver? Are the drivers so different in complexity, size and design that there are no low hanging fruits to be taken from that information? Not that I doubt of their professional integrity, of course : )
Normally the devs work from HW documentation, diagnostic code and information they get from hardware and driver teams (not the actual driver code). Anything pulled directly from the proprietary driver code needs to be reviewed to a greater extent than normal. We basically need to differentiate between "how the chip is programmed" (which we want to release), "how our driver is designed", and "how the chip is designed" (we don't want to release the last two. The practices are constantly evolving, but this is what we do today.
EDIT - I wonder if having to delete and repost rather than edit is artificially inflating my post count ?
In other news, R800 (EG) acceleration will be picked up in Linux Kernel 2.6.48! It means you suckers who bought the cards will wait for another 3 years! haha
In other news, R800 (EG) acceleration will be picked up in Linux Kernel 2.6.48! It means you suckers who bought the cards will wait for another 3 years! haha
It's already in 2.6.35. I'm sure some bugs will be found but AFAIK we are running EXA, Xv and 3D on 2.6.35 today.
You're not wrong though. It will be in 2.6.48 as well
I actually saw quite a big difference with HyperZ on my nasty Radeon IGP 345M. It turned what was otherwise a completely useless piece of crap into something that was almost usable. I was able to switch from 800x600 to 1024x768 in some games like RTCW. Yes, I realise this chip was never meant for games but that's all I had at the time.
Comment