When will Octane render support Cuda 4.1...
I found Arion had released a new 1.6 version based on Cuda, and there will no longer limit the numbers of images.
Will Octane render also fix this?
OctaneRender® 1.025 beta2.57 release [obsolete]
AFAIK cuda 4.1 drivers are still in RC status for a few weeks yet.
I think Refractive will want to get beta 2.6 out though before worrying about 4.1 compatibility issues.
Not sure but I imagine the number of textures is tied to hardware rather than cuda.
I didn't notice anything re textures when looking at the change list recently but....
I think Refractive will want to get beta 2.6 out though before worrying about 4.1 compatibility issues.
Not sure but I imagine the number of textures is tied to hardware rather than cuda.
I didn't notice anything re textures when looking at the change list recently but....
Last edited by pixelrush on Wed Jan 18, 2012 3:24 am, edited 1 time in total.
i7-3820 @4.3Ghz | 24gb | Win7pro-64
GTS 250 display + 2 x GTX 780 cuda| driver 331.65
Octane v1.55
GTS 250 display + 2 x GTX 780 cuda| driver 331.65
Octane v1.55
To Radiance and team,
to be trully phisically correct then, Octane should then implement polarization of light in its light wave spectral transport, so we could then play with "real polarizing" filters on camera in order to tweak diffuse and specular reflectance accordingly to the brdf of the surfaces we're rendering, like we do in real photography (+ tons of cool features that come with it, diffraction, etc...)
I could really use such features...
to be trully phisically correct then, Octane should then implement polarization of light in its light wave spectral transport, so we could then play with "real polarizing" filters on camera in order to tweak diffuse and specular reflectance accordingly to the brdf of the surfaces we're rendering, like we do in real photography (+ tons of cool features that come with it, diffraction, etc...)
I could really use such features...
win7 64|2x [email protected]|48 gb Ram|2xGTX580 3Gb
They don't use the GPU for unlimited textures, objects or materials, they use the CPU.thomaslee1312 wrote:When will Octane render support Cuda 4.1...
I found Arion had released a new 1.6 version based on Cuda, and there will no longer limit the numbers of images.
Will Octane render also fix this?
(look under unlimited objects,
-> http://www.randomcontrol.com/newsletter-2011-12-25)And if you cannot make your scene fit in your GPU memory, you can still render it on the CPU at a very nice speed - Arion's speed.
Windows 7|GT 520M|i5-2410m|4 GB RAM and onboard Intel graphicscard
- thomaslee1312
- Posts: 23
- Joined: Thu Dec 01, 2011 3:09 am
thank you for your replies, I have almost forgot Arion use CPU in the same time. But I tried Arion 1.5, it has the same limitation problem as Octane render.Monex wrote:They don't use the GPU for unlimited textures, objects or materials, they use the CPU.thomaslee1312 wrote:When will Octane render support Cuda 4.1...
I found Arion had released a new 1.6 version based on Cuda, and there will no longer limit the numbers of images.
Will Octane render also fix this?
(look under unlimited objects,-> http://www.randomcontrol.com/newsletter-2011-12-25)And if you cannot make your scene fit in your GPU memory, you can still render it on the CPU at a very nice speed - Arion's speed.
And it is OK now in Arion 1.6, I have tried. So I think it will be the trouble about the Cuda version.
Seems to be working here. Although i had less luck trying to export a scene where one of the objects had polish characters via plugin.roeland wrote:Hello,
Forgot to mention it in the OP, but we also fixed loading of OBJ files with non-ASCII characters in the path. Can anyone confirm this is working?
--
Roeland
Intel Core2Quad 9200 / 4 GB OCZ / Gainward 460 2GB
it seems like the last versions don't release the gpu after the rendering has finished. this prevents the gpu(s) from entering powersave modes, keeping temperatures and power consumption high. is this intended?
„The obvious is that which is never seen until someone expresses it simply ‟
1x i7 2600K @5.0 (Asrock Z77), 16GB, 2x Asus GTX Titan 6GB @1200/3100/6200
2x i7 2600K @4.5 (P8Z68 -V P), 12GB, 1x EVGA GTX 580 3GB @0900/2200/4400
1x i7 2600K @5.0 (Asrock Z77), 16GB, 2x Asus GTX Titan 6GB @1200/3100/6200
2x i7 2600K @4.5 (P8Z68 -V P), 12GB, 1x EVGA GTX 580 3GB @0900/2200/4400
No, we didn't intend to change that, since when have you noticed this?t_3 wrote:it seems like the last versions don't release the gpu after the rendering has finished. this prevents the gpu(s) from entering powersave modes, keeping temperatures and power consumption high. is this intended?
--
Roeland
at least since 2.56; but before you look into it, i'll do some tests with other/older versions and nvida driver versions to make sure, that this is no driver or settings problem...roeland wrote:No, we didn't intend to change that, since when have you noticed this?t_3 wrote:it seems like the last versions don't release the gpu after the rendering has finished. this prevents the gpu(s) from entering powersave modes, keeping temperatures and power consumption high. is this intended?
--
Roeland
„The obvious is that which is never seen until someone expresses it simply ‟
1x i7 2600K @5.0 (Asrock Z77), 16GB, 2x Asus GTX Titan 6GB @1200/3100/6200
2x i7 2600K @4.5 (P8Z68 -V P), 12GB, 1x EVGA GTX 580 3GB @0900/2200/4400
1x i7 2600K @5.0 (Asrock Z77), 16GB, 2x Asus GTX Titan 6GB @1200/3100/6200
2x i7 2600K @4.5 (P8Z68 -V P), 12GB, 1x EVGA GTX 580 3GB @0900/2200/4400