Both dpp-msbayes and msBayes attempt to approximate the posterior of a very parameter rich and highly stochastic model. Furthermore, both methods use relatively little information from the data when doing so (i.e., the sequence alignments are reduced to a small number of simple statistics). Under these conditions we should expect there to be quite a bit of posterior uncertainty about divergence models and the number of divergence events. A lack of posterior uncertainty (i.e., strongly supported results) is likely a red flag that the prior on divergence times is strongly weighting the posterior probabilities of the divergence models (i.e., the strong support is being driven by the prior on the divergence times, and not the data). See the the section about the prior on divergence times for more details.
ABC approximations of posterior probabilities (and associated Bayes factors) can be inaccurate when the summary statistics are insufficient across models (which will always be the case for dpp-msbayes and msBayes analyses) [11]. Thus, ABC model choice should really be treated as a means of exploring your data, rather than a rigorous evaluation of hypotheses.
Given some of the pitfalls discussed above, it is always a good idea to assess the sensitivity of your results to your prior assumptions. Luckily, PyMsBayes makes this very easy for you, because it allows you to analyze data under an arbitrary number of models very easily. See the multi-model tutorial about how to do this.
As shown in [10], [9], and [8], simulation-based analyses provide insight about the time scale at which the method can detect random variation in divergence times from your data. This can be very useful in guiding your interpretation of the results.
See the section about the re-sorting across taxa that is done in msBayes (and is still an option in dpp-msbayes). This is mathematically incorrect, introduces bias, and should not be done.
See the note about the number of divergence models as the number of taxa increase. Once the number of taxa gets above 15 or so, the number of possible models of divergence gets enormous, and might be beyond the capacity of the simple rejection sampling algorithm implemented in dpp-msbayes and msBayes. If you use either method with this many pairs of taxa, you should run multiple replicates, each with large numbers of samples from the prior, to make sure your estimates are stabilizing as the samples increase within each run, and converging to similar values across runs.
Please do not use the default priors. You know your system infinitely better than I do, and thus you should be the one to choose the assumptions about prior uncertainty.