Is the AI considered to be the same person as its human predecessor or a digital twin? Is it a person at all? If an upload is a person, how different do copies of that upload have to be before they're separate persons?
Assuming destructive uploading, the original is dead. How does the copy feel about that? Do they know? Should they be told?
Would the soul be copied over? Is there a soul at all to be copied? While some people might see the debunking of mind-body separation as just another case of science marching on, a great deal of people would find the idea that even their mind is quantifiable to be rather frightening. Or worse, would see those who go through with the upload as less thanhuman, and campaign for a ban of the procedure for it violating human dignity.
Assuming the existence of the soul (or even just assuming the original believes he has one), how does he feel about the prospect that he may not be simply destroyed, but go on to an afterlife (pleasant or unpleasant) while a newly created double takes his place? After all, "he" either stands a 50/50 chance of winding up as the original or the copy, or is always going to be the original. For that matter, is the newborn copy innocent of sin despite his memories of committing them?
Even theorists who don't believe in the soul, per se, often believe in consciousness as a real phenomenon. Would a simulation of a brain experience consciousness any more than a simulation of lungs can be said to actually respirate oxygen? How could an outside observer tell? note Granted that you can't really be sure even that the people around you are conscious like you are; but the fact that they're made of the same stuff as you and function the same way you do provides a pretty strong hint that they are. The fact that the observer probably can't tell arguably makes this consideration more important, not less—since uploadees would be gambling their very selves on the trustworthiness of this tech (and even if it can simulate their consciousness, that doesn't mean it's the same, continued consciousness).
If a scanned mind is an analog recording, the constant and casual re-copying necessary to "travel" electronically would be impossible without corruptingthe data. You could copy yourself into a durable and long-lasting robot body relatively safely, but you could never safely leave it except by physically transplanting the robot's brain. And of course, physical electronic components do wear out (a lot faster than flesh does, at present).
How accurate would the copy be, especially in the early days of the technology? If the flaws are significant but not immediately obvious, how many people might undergo the procedure before the problems are noticed? And if you know about the flaws ahead of time, how much of your personality or consciousness are you willing to throw away or see changed beyond your control for a type of immortality?
If you have concerns about the trustworthiness of the process, if everyone you know is doing it? Conversely, if you're a true believer in the process, what if society condemns it?
Can the computer provide a good enough simulation of human sensory input to keep you from going mad? Even a brief period spent in a sensory deprivation tank can have terrible effects on the mind, so one can imagine what complete absence of a physical body might do.
A man converted into software has all the vulnerabilities of software. He can very likely be hacked, duplicated, or edited against his will. For better or for worse, the human mind is currently relatively impregnable. Do you really want to be rendered no more unique than a google search image, and more malleable than putty in the hands of others? Do you want to wake up one day to find that you're an illegal copy of yourself, being treated as a toy by a hacker? Would you necessarily own the copyright to yourself? If such a copyright even existed at all (since many consider copyright unenforceable andundesirable in the digital age), would the agency that uploaded you own it? How can the law provide any protection to a citizen who can be duplicated (and his duplicate used andabusedhowever the criminal wishes) as easily as copying a computer file? And every time such a copy is produced, "you" stand a 50/50 chance of being that twin. If a virtual world makes a synthetic heaven possible, it likewise makes synthetic hells possible, and the latter may be far easier to produce (either accidentally or deliberately).
In a world where uniqueness exists, at best, as a legal courtesy, mightn't human life come to be seen as fundamentally less valuable?
Who owns the computer or computers that your virtual self would be running on? Are they under any obligation to keep running your program? If your program is not being run, is that the same as being dead? Asleep? In suspended animation? And if you don't like what they're doing or are planning to do with you, could you class any attempts to stop them as 'self-defence'? How do your rights as a human being translate to a computer simulation?
What if you changed your mind? What if you don't like it? Would you be allowed to simply turn yourself off?
Would a human program be able to perceive anything outside of their device? Or could they percieve anything at all, beyond the digital data fed directly to them?