To be honest the line between the two is almost gone nowadays and there are processors that can be classified as both (AD Blackfin for instance).
Generally speaking:
Microcontrollers are integer math processors with an interrupt sub system. Some may have hardware multiplication units, some don't, etc. Point is they are designed for simple math, and mostly to control other devices.
DSPs are processors optimized for streaming signal processing. They often have special instructions that speed common tasks such as multiply-accumulate in a single instruction. They also often have other vector or SIMD instructions. Historically they weren't interrupt based systems and operated with non-standard memory systems optimized for their purpose making them more difficult to program. They were usually designed to operate in one big loop processing a data stream. DSP's can be designed as integer, fixed point or floating point processors.
Historically if you wanted to process audio streams, video streams, do fast motor control, anything that required processing a stream of data at high speed you would look to a DSP.
If you wanted to control some buttons, measure a temperature, run a character LCD, control other ICs which are processing things, you'd use a microcontroller.
Today, you mostly find general purpose microcontroller type processors with either built in DSP-like instructions or with on chip co-processors to deal with streaming data or other DSP operations. You don't see pure DSP's used much anymore except in specific industries.
The processor market is much broader and more blurry than it used to be. For instance i hardly consider a ARM cortex-A8 SoC a micro-controller but it probably fits the standard definition, especially in a PoP package.
EDIT: Figured i'd add a bit to explain when/where i've used DSPs even in the days of application processors.
A recent product i designed was doing audio processing with X channels of input and X channels of output per 'zone'. The intended use for the product meant that it would often times sit there doing its thing, processing the audio channels for years without anyone touching it. The audio processing consisted of various acoustical filters and functions. The system also was "hot plugable" with the ability to add some number of independent 'zones' all in one box. It was a total of 3 PCB designs (mainboard, a backplane and a plug in module) and the backplane supported 4 plug in modules. Quite a fun project as i was doing it solo, i got to do the system design, schematic, PCB layout and firmware.
Now i could have done the entire thing with an single bulky ARM core, i only needed about 50MIPS of DSP work on 24bit fixed point numbers per zone. But because i knew this system would operate for an extremely long time and knew it was critical that it never click or pop or anything like that. I chose to implement it with a low power DSP per zone and a single PIC microcontroller that played the system management role. This way even if one of the uC functions crashed, maybe a DDOS attack on its Ethernet port, the DSP would happily just keep chugging away and its likely no one would ever know.
So the microcontroller played the role of running the 2 line character LCD, some buttons, temperature monitoring and fan control (there were also some fairly high power audio amplifiers on each board) and even served an AJAX style web page via ethernet. It also managed the DSPs via a serial connection.
So thats a situation where even in the days where i could have used a single ARM core to do everything, the design dictated a dedicated signal processing IC.
Other areas where i've run into DSPs:
*High End audio - Very high end receivers and concert quality mixing and processing gear
*Radar Processing - I've also used ARM cores for this in low end apps.
*Sonar Processing
*Real time computer vision
For the most part, the low and mid ends of the audio/video/similar space have been taken over by application processors which combine a general purpose CPU with co-proc offload engines for various applications.
The bootloader lets you reprogram the PIC from the USB port, taking advantage of the ability of the controller to write into its own program memory. Usually, it checks to see if some criteria is met, like some bit being set high, before entering the programming mode. It will shift your program into higher memory space to accommodate the boot loading protocol at the normal start vector.
This is the way to go if you see a need for something like an end-user updating the firmware on your device without special programming hardware.
Serial port boot loaders are a bit easier to learn, as you don't have to deal with the USB stack.
Best Answer
Some 18F nomenclature but in German, so maybe use google translate?
All the PIC32's have a 'part number decoder page' for example page 328 "PRODUCT IDENTIFICATION SYSTEM" in PIC32MX1XX/2XX
or page 656 "PRODUCT IDENTIFICATION SYSTEM" in PIC32MZ Embedded Connectivity (EC) Family
AFAICT, all PIC datasheets have similar part-number-encode/decode pages. However, I'm only interested in PIC32.
Wikipedia PIC microcontroller/Family core architectural differences
That divides the families by instruction size, and gives a useful summary of the 12bit, 14bit, enhanced 14bit, 16bit, 16bit dsPIC, PIC32 devices. Like many other microcontrollers, the 'natural' data size and instruction size can be different.
Edit:
I'm not a PIC24 person. Looking at the datasheets for PIC24FJ256GB210 and PIC24FJ256GB110, PIC24FJ256GB210 has 96kB RAM vs 16kB RAM for PIC24FJ256GB110
Typing the part numbers into "Search Microchip" at Microchip usually gets to a brief summary of the part, and links to more information.
For example PIC24FJ128GA010
Typing the part numbers into "Search Data Sheets" at Microchip will get to the device datasheet directly. Data Sheets usually has a summary of each family part on the first real page (page 3?). The table is often good enough to compare across similar families, and identify the differences.