Summary: | cal3d 16bit-indicies cause eternal lands to segfault | ||
---|---|---|---|
Product: | Gentoo Linux | Reporter: | Branko Majic <branko> |
Component: | [OLD] Games | Assignee: | Richard Freeman <rich0> |
Status: | RESOLVED FIXED | ||
Severity: | normal | CC: | games, rich0, uberlord |
Priority: | High | ||
Version: | unspecified | ||
Hardware: | AMD64 | ||
OS: | Linux | ||
Whiteboard: | |||
Package list: | Runtime testing required: | --- | |
Attachments: | Backtrace with (I hope) some useful information |
Description
Branko Majic
2007-05-09 20:17:34 UTC
I'm not seeing this on my amd64. Can you provide attach backtrace please? http://www.gentoo.org/proj/en/qa/backtraces.xml My backtrace won't provide any useful information - the system's binaries are stripped of all debugging symbols (except that one time when I tried compiling EL with debug symbols, but it doesn't help either). Now, I'm using gcc-3.4.6-r2, so I'm thinking if that may be the problem? Seeing that other people have no such problems, I'm beginning to think that it's only occurring on my machine. Btw, EL 1.3.3 stopped working after I updated the world at some point. I have no idea what to do (except to switch to new gcc 4.x, but it will take a lot of time to recompile the system). (In reply to comment #2) > My backtrace won't provide any useful information - the system's binaries are > stripped of all debugging symbols (except that one time when I tried compiling > EL with debug symbols, but it doesn't help either). Now, I'm using > gcc-3.4.6-r2, so I'm thinking if that may be the problem? Seeing that other > people have no such problems, I'm beginning to think that it's only occurring > on my machine. Btw, EL 1.3.3 stopped working after I updated the world at some > point. I have no idea what to do (except to switch to new gcc 4.x, but it will > take a lot of time to recompile the system). > You need to follow th GCC upgrade guide. http://www.gentoo.org/doc/en/gcc-upgrading.xml#upgrade-general Re-open when you have done this and still have this error I've upgraded my gcc, and the problem persists (hey, at least I got my gcc upgraded, eh?). I'll try to rebuild some of the game's dependencies with debugging information, and then run it through gdb. Created attachment 119118 [details]
Backtrace with (I hope) some useful information
I'm not sure if I got all of this right, I hope I did. Unfortunately, I didn't recompile my glibc with debugging symbols. If it would be more useful to do that, let me know.
OK that's showing an error in the OpenGL part shipped by nVidia. Do other 3D applications work fine? For example, we have ut2004-demo in portage, does that work? How about 3D OpenGL screensavers? As far as I can tell, other games work just fine. I've tried playing X2: The Threat - it works. Sauerbraten - works. Savage - works, bzflag (the only game listed here which is not distributed in binary, I think) - works, several games played through wine work as well. I've also tried using older nvidia drivers (the non-masked ones), but as I said - it still segfaults. When did you last update your packages? Ah, sorry, I'd also like to add that non-game OpenGL applications work as well (I use a screensaver which uses OpenGL), Beryl included. Have you tried running EL without compiz and Beryl? Also try disabling compostie, XGL and AIGLX if you have them enabled. As to the packages, my system runs a full ~amd64 - which means all the latest libraries. Neither Compiz nor Beryl are running on my machine all the time (I mostly use Beryl for showing-off due to the ugly bug in nvidia-drivers). Hm... I'm using mostly stable branch (a package here and there is from ~amd64), the only direct link to this problem might be nvidia-drivers (I use version 1.0.9755-r1, which is marked as testing), but as I said earlier, I've tried using the stable version to no avail. It might be interesting to have someone who runs mostly stable amd64 to test this thing as well (apart from myself). I'm pretty sure that some of the underlying libraries are messing things up, maybe the newer libs you have work better? Btw, are you sure the segfault occurs in nvidia drivers, since I'm not sure that my trackeback is good enough? I think I'll try to get X up without some specific options in xorg.conf that allow the usage of Beryl. I have (finally) identified the problem. The problem was the '16bit-indices' USE flag in media-libs/cal3d-0.11.0-r1. I've tried recompiling this library without this flag, and the problem is gone. Then, just in case, I compiled it again with this USE flag, problem is back. Tomorrow I'll do a check on x86 machine in order to see if it's arch-specific (so far I've encountered several problems regarding the sizes of various integer types on amd64). Since you're running testing branch, it would be nice if you could verify these findings (just to make sure). Good catch! I've added this check to the ebuild, thanks. re-opening as cal3d should not cause this crash, but it is a cal3d bug No clue why this is assigned to python. Assigning back to maintainer. Looking at eternal-lands it seems it redefine CalIndex, that is the indices that is affected by the 16bit-indices option. The bug should be on that side. Forcing the usage of cal3d [-16bit_indices] is one way to fix it, and it is done. |