Jump to content

vozka

Member
  • Posts

    119
  • Joined

  • Last visited

  • Days Won

    1

Everything posted by vozka

  1. Added some chipped paint to the painted version. I'm the type of person who hates doing things by hand so I spent about 5x more time learning how to do it procedurally (surprisingly there are a few catches), but I'm pretty happy with it and I think that makes this one worn enough.
  2. If one out of ten people notice, that's good enough for me, the only issue is if I'll be able to make it low-poly enough to not strain the engine. I tried making it low-poly and bake normals in a quick & dirty way, so there are errors and I'll have to try a couple times, but so far I got the whole handle down to 170 tris. Seems like with a shadow mesh, which I guess could be half that, it would be good enough.
  3. Grunge and distressing is definitely something I have some trouble with, but you're right. Regarding the inlay, good point - I didn't really do that because this is still PBR with metalness and everything and the materials will have to be done slightly differently for TDM anyway. While this is flattering, it's not that difficult to get a semi-photorealistic render in a path-tracing renderer using physically based materials and models with a virtually unlimited number of polygons. Adapting that into the game is another matter I created two variants. First is a black paint (again, no distressing yet) that imo looks a bit better, but if I recall correctly it's possible to have models with more than one set of materials, so it would be best to have both. Second is a new handle what apparently was semi-common in the victorian times and I thought it was pretty funny. No idea how it will look after low-poly retopo and baking though.
  4. Awesome, that would be great. Regarding the separate models, that's what I assumed, so that's how I did it. I'll work on it a bit more first.
  5. Well if this is deemed good enough, I'm probably going to need somebody to help with the interactivity anyway - door opening and the lock/moving the handle. I'm not a map maker and I'm not sure I'll have the time to learn that (it's harder to justify than practicing modelling, which I need for other things too). Providing a version without the keyhole would not be a problem for me of course. I have an alternative material and handle in mind as well, it would be nice to have more than one option.
  6. Attempting to make a model of a safe for fun and practice. So far it's just hi-poly with PBR materials. My modelling skills are quite narrow and this is outside their scope, so I'm looking for criticism, mostly: - Is the overall style acceptable, regarding apparent time period etc? (apart from the text in the logo, that's a placeholder and I imagine I'd use some decorative victorian font plus some actual believable names) - Would the quality be good enough like this? - Do you have suggestions about how to improve it?
  7. Yes. Not as good as Dall-E at this moment (but will improve soon, several improved versions are already in works or done and will be gradually released), but free and open-source.
  8. An article about using Stable Diffusion for image compression - storing the data necessary for the neural network to recreate the image in a way that would be more efficient than using standard image compression algorithms like jpeg or webp: https://matthias-buehlmann.medium.com/stable-diffusion-based-image-compresssion-6f1f0a399202 It's just a proof of concept since the quality of the model is not there yet (and the speed would be impractical), it can't do images with several faces very well for example. But it's very interesting because it has completely different tradeoffs than normal lossy compression algorithms - it's not blocky, it doesn't produce blur or color bleed, in fact quite often it actually keeps the overall character of the original image almost perfectly, down to grain levels of a photograph. But it changes the content of the image, which standard compression does not do.
  9. Also most missions have so few ratings that the average is kind of random. If you're just looking for some order in which to play the missions than this is not a bad one, but it's not really worst the best and don't take the ratings too seriously.
  10. This is already possible to a degree. It takes two things: first you use the language model of the AI to "interrogate" the image and find what text description would lead to generate something close to it. Then you basically use the diffusion algorithm which normally generates images from noise in reverse, to gradually generate noise from an image. This gives you the seed and description, and the result is often very close to the original image. I was playing with it yesterday because it then allows you to change some details without changing the whole image, and that allows you to take the portraits of your women friends and transform them into being old and unattractive, which I found all of them universally appreciate and find kind and funny. The early implementation in Stable Diffusion is just a test and it's quite imperfect (it produces noisy images with overblown colours among other things), plus it's limited by the quality of normal SD image generation, but as a proof of concept it obviously works and will probably get better soon. However I don't see this happening any time soon. The AIs work with bitmaps, are trained on a specific image size (with Stable Diffusion it's 512 x 512 px at the moment) and making anything significantly smaller or larger completely breaks the process. Various AI upsampling algorithms exist, but they never work as well as straight up generating the image in the resolution that the neural net was optimized for. And I don't know about any practical solutions to this yet.
  11. This is offtopic, but you just reminded me of a great memory. I was always into underground music culture, mostly metal as a teenager, but later also electronic music. A few months after I moved to a different city for university I discovered a tiny underground (literally) private bar, basically a rave speakeasy. No license, entrance with a camera, located in what used to be a wine cellar and a later a PC gaming arcade (which left very strong air conditioning there, so the weed only smelled outside, not inside). Each weekday it was open from about 6-7 pm to about 4 am, and almost every time there were one or more DJs playing more or less underground electronic music, for free, just for fun. Sometimes doing live broadcasts over the internet, before that became a common thing. I still remember the first time I came there, walked down a narrow set of stairs and entered the dark main room with a couple tables with benches and about 10 different CRT monitors placed in random places (including under the benches), all playing Electric Sheep animations. At the time it felt like something from The Matrix. The place is still kind of operating many years later, but it got handed over to guys who care more about getting drunk everyday than about music or technology, so the magic is long gone.
  12. Neural networks can and do forget things during training, and their memory is very imprecise. And we're not entirely sure why that is, as far as I know. But it's one of the basic principles without which the whole concept wouldn't work. It is true that a neural network cannot have a sudden flash of insight, or more precisely it cannot realize that it just had a sudden flash of insight. Artificial general intelligence does not yet exist and is probably a couple decades away, and specialized AI is not self-aware. It can certainly act as a flash of insight for a human that uses it to create something unexpected though.
  13. You make some good points, but I personally think it's similarly naive to think that we cannot eventually get to a point, where an artificial intelligence produces results so similar to a human in some areas, that the fact that it functions differently becomes irrelevant. In fact I'd say that current neural networks are a big step forward both in terms of the results they produce and in terms of how similarly they work to a brain. We know that artificial "neurons" are a distant abstraction of how a brain works, but a neural net is not just a normal mathematical algorithm. Yes, it's "just" matrix multiplications, but at the same time the complexity of the computation is so far beyond what any human could analyze or design that it's a blackbox. That in particular is similar to a real brain. The output of a neural network is deterministic just like with a "normal" algorithm - two identical inputs give you two identical outputs. With the case of image generation algorithms part of the input is noise (in the shape of the final image), which dictates how will the final image look. Change the noise very slightly and the output can be completely different. Since we want the results to be reproducible, we use a pseudorandom noise generator initialized by a seed number. Same seed will produce the same noise pattern and therefore the same resulting image. What if we instead use a noise generator that is truly stochastic? We are able to do that (using specialized hardware). The algorithm itself would still be deterministic, but it would be unpredictable because a part of the input would be unpredictable. Would that be "organic" enough? And what if we stop computing the neural network on a GPU and instead use a chip that does analog matrix multiplication (already exists, although not powerful enough for this particular usecase, yet), which brings a stochastic element into the computation itself? The computation itself is no longer deterministic. You know what will roughly happen, but cannot predict the exact result even if you know the input. Would that be organic enough? It is certainly a very different thing now from how any normal computations work.
  14. This will undoubtedly happen with large parts of digital art. But there's a ton of art in the physical world that is not in danger. Even oil paintings - you can print digital art on canvas, but it won't be able to mimic some specific painting techniques that create a physical structure on the canvas, or specific types of pigment. Or various types of combined media. Then there are other even more physical media like sculptures, that's an infinite world. Honestly from the various art school graduates/teachers that I know none of them are threatened by this because none of them ever did normal digital art like anything an AI can produce.
  15. That's the beauty of it being open-source. An unofficial AMD port already exists (although I have no idea how well it works) and people did several optimizations that reduced the RAM necessary to run it down to about a half (with almost the same computation speed, less than half with reduced speed). My results are just a proof of concept and not very useful for game assets, but I was pretty happy with them. I made a very simple 3D render with unnaturally clean PBR materials, and then added some dust and dirt on them using Stable Diffusion. I had to overlay some black noise texture over the clean materials because the AI needs something dirt-like to work with, it will not generate dirt on something that is completely clean unless you let it deviate from the base imagea lot, and in that case it tends to change the shape of items and distort things. I also had to render the bench and garbage bin separately (but keep their shadows in), the AI tended to change them too much. The images are 832 x 832 px in size because that's the maximum that currently fits into my 6 GB of VRAM (I'm not using the latest optimizations), but when doing something like generating noise (which doesn't need much continuity), you can easily just split a large render into smaller tiles and do it tile by tile. Before AI dirt: After AI dirt:
  16. I do not have permission to say names, but the concept artists I mentioned work for a mid-sized company and they use Midjourney for scifi themed concept art. The results surprisingly are not generic looking at all if you learn to use it, but it's true that the "text to image" workflow does not give you enough control to straight up generate assets (and the quality is not good enough either). Midjourney is a paid proprietary service, but they specialize in their results being slightly less photorealistic but more creative and artistic. But the important thing is that the "image to image" workflow exists. Normally with text to image you describe what you want, the the neural net starts with noise and gradually rearranges it into something that resembles the description. With "image to image" you start with some actual image that you feed to the neural net instead, and then have it transformed to something that fits the description. You can change how far the program is allowed to deviate from the base image. This gives you a lot more control and it could potentially be used to generate textures for example. Some guy just posted an example on Reddit. He drew a rough scheme of a "scifi wall" using basic shapes, overlayed a random scratched metal texture and had SD generate several versions of a scifi wall from that. Then he edited his favorite result a bit and fed it back into SD, generating the final version. This is how it looked: I used the same workflow when I was adding grime to a final render (so not straight to a texture) and it worked well too.
  17. I assume all of you have at least heard about it already or seen some pictures. Neural networks that generate decent looking images from a text prompt and do some other things (like modifying existing images) are becoming mainstream. They cannot generate photorealistic hi-res images yet, but the development pace is incredibly fast, so they will probably get reasonably close within a few years. The big thing that happened recently is that Stability AI released their implementation of such neural network, Stable Diffusion, for free as open-source. The quality is not the best, but it's good enough and all the other ones are behind a paywall, waitlist or just a closed source service and not good enough. You need a GPU (preferably Nvidia) with at least 4 GB of RAM to run it, so nothing special. The potential implications to society are big (and not necessarily positive), but honestly I don't really want to start the discussion because the internet is already full of it and so many people are surprisingly emotional about it. Go to Twitter if you want to see (you don't). I'm more interested in what you think about using it in game development. I know that in some commercial studios it's already being used by concept artists, because it's firstly a reasonably efficient inspiration machine and secondly the images it produces are not licensed (and it doesn't matter that much in pre-production anyway), so if you manage to create something good, you can just paint over it to finetune and use it. But it can do other things. It can do inpainting, which can be used to remove seams or blemishes in phototextures. It can also generate seamless textures (there's a switch that causes every generated image to be periodic). There are no examples of people using them in games or renders yet because that feature is less than a week old. Personally I tried using it to add dirt and imperfections to a render that was too clean, and that was a success as well. Adding dirt to a render is not exactly useful in most of games, but I bet it could be adapted to add some grime to game textures as well. Any ideas? Have you tried it already?
  18. vozka

    Monomyth

    Reminds me of Arx Fatalis or Ultima Underworld. I wasn't looking forward to it, but I am now.
  19. Really nice mission! I haven't played any of your old ones so I had no idea what to expect and it was a very pleasant surprise. Tight and great looking, the environment was really well made. Appreciated some of the new (or less used) soundscapes as well, I think it's a big part of the overall atmosphere and while pretty much all of those used in TDM are good, the effect gets lessened when you hear them all the time. Looking forward to you next work!
  20. I encountered the bow crash again, in a different mission. It's in Who Watches The Watcher?, it happened again when shooting an NPC with a broad arrow (a different NPC than in Hazard pay, a guard, so it's not zombie-related). I noticed one new thing - the bow shot that crashes the game has broken animation. Something happens when "loading" the bow, the (broad) arrow seems to vanish and maybe the bow is at a weird angle as well - I'm not sure since it happens pretty fast and I only saw twice. Then I think the arrow release still works, and as soon as the arrow (dunno if it's visible by then or not) hits the enemy, the game immediately crashes. I am only able to reproduce it randomly. I have a save created shortly before the crash happened and I manage to reproduce it once in 2 tries, and then in the next 15-20 tries I was not successful again, so I gave up. I only managed to reproduce with broad arrows, not with water, gas or fire arrows that I had on me, but that may have been by chance.
  21. Very enjoyable, thanks for this mission. I'm just watching the Snowpiercer series (which is pretty good by the way), so this fits right in thematically. Great idea, and it seems to me that a train mission is like the opposite of sewer missions - everybody hates those and they're almost always bad, whereas when a game dares to do a train mission, it's almost always good. I would have liked if it was longer simply because I was having fun. Can't have it all though and compact missions are also nice. Getting enough loot was just fine for medium difficulty, but I honestly have no idea where the rest of it could have been hidden, so good job on that too.
  22. vozka

    Free games

    A bit different topic since his games are free all the time and not just during giveaway, but do you know of Increpare Games? He's producing a staggering amount of puzzle games, some of which are really smart and really good. Others just funny and stupid. Often the first goal of the game is to understand how it works and what's the goal. Gestalt OS for example is one of the best puzzle games I've played.
  23. Nope, I don't like downloading codec packs because I remember that they used to create messy situations in the system by containing outdated codecs, unneeded codecs that conflicted with other apps etc. That was about 15 years ago and it's possible the situation is entirely different, but for me it stuck (and up until now I haven't needed to install any codecs separately for about a decade). There's a google webp codec and a windows store webp codec, I used the google one.
  24. Well, yeah, the problem of slow adoption is precisely what I'm complaining about. This is kind of a circular argument - if nobody adopts it, nobody is going to use it and it's going to have little support. I do realize that this is a difficult problem to solve because companies have to basically volunteer to "passively" support it and hope it brings them advantages in the long term. There are other aspects that I omitted as well, like for example HEAACv2 being slightly more demanding to decode (which is also the reason why Bluetooth historically used poor SBC encoding), but those are largely moot nowadays as well (even integrated bluetooth chips and cheapest SoCs can decode pretty much any audio format with no issue. Your point regarding storage and network costs does not cover the whole situation, I see three issues with it. Firstly mobile internet is still slow in many places in the world and it's often limited. Where I live it's difficult to get more than 5 GB monthly limit for a reasonable price. Secondly most websites are bloated as hell and any way to speed up their loading helps. This was the main reason for webp creation and adoption. Thirdly for services like music and video streaming bandwidth is expensive and the difference between using 160 kbps and 96 kbps for the same quality can save a lot of money. That was the reason why Youtube and Soundcloud switched to more modern codecs. And not even that, Netflix in some apps already switched to one of webp successors for movie thumbnails, because in such a large scale even serving the thumbnails in a decent image quality costs a lot of money. Btw in my case webp was certainly not useless, it was just an annoying situation. I use it as backup for my own use (seems to be about 1/5 of png size so far for photo-like files), so I just downloaded a codec from google to get explorer thumbnails and I don't need other people to support it. Although the fact that Facebook messenger converts webp files to... static gifs of all things, is kind of annoying (and funny).
  25. Since this thread is already silly, why not derail it a bit: The main thing I took from this whole ordeal is that it's annoying how long it takes to adopt new and better formats. There are virtually no reasons to use .gif apart from compatibility with ancient systems. Jpeg from an actual image editor instead of mspaint being much better is one thing, but there's also webp, heic, avif, jpeg xl and others, all better than their lossy or lossless predecessors and most have free implementations. Except almost nobody uses them because the adoption is glacially slow. Surprisingly Chrome and Firefox are fast, but everyone else isn't, Apple traditionally being the worst. I was pleasantly surprised to find that Blender now supports webp output for rendering, which in its lossless version is straight up better than png or tiff, only to find out that Win10 doesn't support it natively and I have to search the web for a codec to get thumbnails in explorer. It's the same with music files. We've had codes better than mp3, standard aac or ogg vorbis for almost 2 decades now, yet these three are still the most used codecs for streaming music. And when a platform decides to use something else, it's often a step down in quality because they push down the bitrates too much, as with youtube in the last couple years.
×
×
  • Create New...