Would make sense, both the delay and range of fuckups from and including decision making to hand action are all but eliminated by something like this. So on the one hand it will do wonders for anyone with dyspraxia, but on the other it ups the ante on competitive input peripherals.
If it's allowed, yes. But it's going to be so revolutionary it will either need banned or its own category made.
tl;dr this is unironically going to lead to real world Deckers from Shadowrun.
Response time is going to drop significantly with this. The thought of doing something is going to be the input command, not the current thought which then needs an appropriate neural command sent to the hands to perform various actions that provide the input, those actions as I mention in the other post being subject to fuckups and delays.
RSI, twitches, itches, outside interference, shitty controllers, keyboards, mice, WiFi, and more, all replaced with the integrity of signal interpretation, which will have to be paramount in the first place, and transmission.
It's entirely possible this results in player performance leap frogging current game challenges, too, because right now games are designed with the limits of both user and hardware limits in mind.
At best we can learn to input/type faster alongside using additional input devices but we are still limited by physical features of both the user and the devices present. A mouse and keyboard means you can do several things at once, but we still only have 2 hands. So we can move in a direction, jump, turn, interact, and do so all at once or in various combinations. But we will still be limited by the game design just how much we can input those commands.
With a 20+ year old ps2 controller a user can use the left analog stick to move in any direction along the horizontal axis. But even if a second analog stick was provided that permitted true 3D movement within an environment, air or water, we don't have the dexterity for that level of finesse. A direct neural input however would because without needing to calculate how hard a button needs pressed or how much a second analog stick needs turned a human brain can interpret spatial data and create a vector to move along. Every single game that exists right now which offers superpowered flight is going to be redundant overnight with this type of input device.
Didn't the first guy to get it come out and say it made him an aim bot? I think he was saying he had an unfair advantage
Would make sense, both the delay and range of fuckups from and including decision making to hand action are all but eliminated by something like this. So on the one hand it will do wonders for anyone with dyspraxia, but on the other it ups the ante on competitive input peripherals.
So the next game competition will basically be cyborgs?
If it's allowed, yes. But it's going to be so revolutionary it will either need banned or its own category made.
tl;dr this is unironically going to lead to real world Deckers from Shadowrun.
Response time is going to drop significantly with this. The thought of doing something is going to be the input command, not the current thought which then needs an appropriate neural command sent to the hands to perform various actions that provide the input, those actions as I mention in the other post being subject to fuckups and delays.
RSI, twitches, itches, outside interference, shitty controllers, keyboards, mice, WiFi, and more, all replaced with the integrity of signal interpretation, which will have to be paramount in the first place, and transmission.
It's entirely possible this results in player performance leap frogging current game challenges, too, because right now games are designed with the limits of both user and hardware limits in mind.
At best we can learn to input/type faster alongside using additional input devices but we are still limited by physical features of both the user and the devices present. A mouse and keyboard means you can do several things at once, but we still only have 2 hands. So we can move in a direction, jump, turn, interact, and do so all at once or in various combinations. But we will still be limited by the game design just how much we can input those commands.
With a 20+ year old ps2 controller a user can use the left analog stick to move in any direction along the horizontal axis. But even if a second analog stick was provided that permitted true 3D movement within an environment, air or water, we don't have the dexterity for that level of finesse. A direct neural input however would because without needing to calculate how hard a button needs pressed or how much a second analog stick needs turned a human brain can interpret spatial data and create a vector to move along. Every single game that exists right now which offers superpowered flight is going to be redundant overnight with this type of input device.
I look forward to the non wired game inputs.
Like a dude boing a chick at the Olympics?
I see a whole new level of dissing coming to a lobby near you soon :)
Judging by the footage in OP, that was probably a joke reported improperly by the media, if it happened at all.
I have a receipt that the first patient called himself an aimbot.
https://archive.ph/0MPcM